diff --git a/.gitattributes b/.gitattributes index 22d43eccc054154fdd49daa00f949a70d8c4c2fe..b255de9dd7d1ad2d650c0205fed4acadfeb0b660 100644 --- a/.gitattributes +++ b/.gitattributes @@ -679,3 +679,10 @@ runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lf runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json filter=lfs diff=lfs merge=lfs -text runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json filter=lfs diff=lfs merge=lfs -text runs/mask15-l2r50-fulle-lm/checkpoint-7200/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eab74536f5292607bb166ad567c81f2a71d212e5 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7456ca1104a53851cb33f03ad5729d8c58e31c480e5e1fd319c864031c1d4024 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3b99eafc944be08e97bfa7b25ec548e04f179dfa --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86273e83eba0e7d0a005a4d73b6f3b0c73fbe7a5f68231fced6a0b61d3ad406b +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..f953b3d2601997ff26348cfe65bdf6f072da1241 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62b735ad70e5179210c3231e43080bd2451d7d358b5efe5db5997cbd75fd394d +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..90eeeb9606172a80974fd4d514e5c336ea10bede --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e39f26de4d91b53c07b4529fc28b0d8db87f95f5cd65b50c9d4f7b36f41aac1 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..41b5a0aaa0603136050870ac83e05be64379af70 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d09a5602d03fd240123f234cec1b33801653989554e35eede4963f7b24092f96 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..5fa2d1f78ea851b45326b8bc87837827bc74fe42 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:751dd23154d8ff44e1735c7b0fd261bc89701be5ab5e929dbb27efa8d599b231 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..630fbc73d70608487940c0c34e65caee42415237 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a67914066d1aebda73626cb9d7d4cbc46ee7df332e5ae85a86cc87be8e519c8 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..122680c53242df36b46a84d812209adda9132756 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json @@ -0,0 +1,8218 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 16.0, + "eval_steps": 500, + "global_step": 1152, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 4.9113813781315584e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a95eb51a71cd11b85f9e052497cbed691f196824 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a7ef37b8c5a1dc5e759d2999fc4d9b087b2eaa5bc0fb5009e3ea232ff9a271c +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..42d3088c631310994a23360d6e906b1f73b7bc60 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd15e0e248cb1e32f608a1690428b2e487cf4635fdee6ef078225c433a16402b +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..fa6594ba5f910ea66238338f849a8e953e567e1a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff0d1c60dbaf626d9ffd2cf5f49d2a345d6f5acba8013a96c4d1b9b99d743250 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..e062f681cfc898c2151a607d123b890be40e76d8 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92360cbd2f45549906da13b2089de2060a693f31beb68deff404d66ba4058b9f +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..cbf904db0a94ff0866429f1dee2aad9896198c54 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43ae0c839fafd8f4cae880ebeae7081f01518492971a5a9301e81c43373bafae +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..322e7e0906e04eea8aa972bd0c5855ca1ca3fa97 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a970baceb23ab09484fef82bc4cb2e61052c409e73fd0fd60914013c019d38d6 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0d34fe10b069b50b5fecbd4a0affbd8f653b7b9b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a84bbc84bb0dd489a2d72fd441cc0b8482e9e06d7cde3ac44b50a389883ce86 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..4e979d33cc004ab839d678dd86839e9af319496f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json @@ -0,0 +1,10266 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 20.0, + "eval_steps": 500, + "global_step": 1440, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6.139226722664448e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-1728/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-1728/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5b232e6b5d30199e8cae2325e41456d50419a6ca --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f24d233851bba30c3d5184941fd624e9c0099141a75a2ab4bc1f1d4f6ce606e +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1728/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..c0378d08c4b5c930a5bf4d85c245e568ad70b9cd --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f6584b3817173853f997d68041230ab13a91932ef5f6e3e446e4c989d7f672f +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..441c22c7739a4a330ab7b7bc71a1ac30ae88d070 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d7091aa49cc705439119e0681a55221108bcd330a7cddbab2f08ea063d2bc57 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..3db45464d5c2e27b3ea96dc60d48c27b1d6456be --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3628985308248fe16453703204feca526901cbada0027336a1d40b0a847018d9 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..723c5a12484ed7058823a51ed4d22b3a8e90093d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b345f134315d19b00d8eca0e14d68627acead75dcd96900a83ea3573c76bc71 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..c213a79c2b53b3b860b2a41aa1f723a98c3e9d28 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fd22b557bc4e909e131aafc4f0279bca66927adbdaae3b726b651f76b7208d6 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-1728/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..baba916c0cec91178026a4cf9d1f610779561fa9 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6a6171dda32af3a9caef592bd2cfcdef1224bfed4ddda14026d355f2f43e8f1 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-1728/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..236c964fbfdd98802b3f174c21121bb546dc11df --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/trainer_state.json @@ -0,0 +1,12314 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 24.0, + "eval_steps": 500, + "global_step": 1728, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7.367072067197338e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-1728/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-1728/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-1728/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2016/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2016/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..609c1b93849f9cb071c4d3b6968fa93a69562b4f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e00c75ff786934715e62f92203ed4883abf95a65b5ffb07ef440b9fb7088a976 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2016/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..622ef6b55f2312aa6e7055620c9d6684fbea5193 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2ef370b20e344a29bc7d95c4e6a4be06765f0073271e325dee5354c06500cc7 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..fe1a4bdbccb862d60c07f9b4d54d6913589aa4a8 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8dcc5baf9a16093438252563827f7b6a020c20d8bdd59a2434185bca6b10d55 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..454a4ec443a59d30815357e097257bfe87492944 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d1f4cbabe5810040153ebf881570d58f8d5b569da22e6703989d9622df57a2b +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..7281baf097ae7539af2086a75ab2f4aabf11a454 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:467348aa2353e1b0a97dac51f9a54d62cebca24058072cc53462eca16b886251 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..37625fad4a311eebea8762bceb0d3b43cd6fbd3d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:958522b68e2cb9ce69ad44e4ae6a2ea4c6ce0bc22433642dce3b33306f5be3fc +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2016/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..b92e75de04edb55fd4d6a1778589870a5bc904d0 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71f52a3ae3412b2053eb61b18dd446a6b3125f82036fb38a407b55123d55fdf7 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2016/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..09b7a92a89f2dfae040dd44f3d79c6234258eeda --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/trainer_state.json @@ -0,0 +1,14362 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 28.0, + "eval_steps": 500, + "global_step": 2016, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 8.594917411730227e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2016/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2016/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2016/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2304/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2304/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4337a74bfb72bc6067e870fcd49a307876c60026 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:612be7549a62726fad1ba1b692e96ff4def8b87fe5608f4921acfb3519bca075 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2304/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..67837922fac0e574d863ddd1e699f0129b800285 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c82e59d4958422cd14ffb582877b08f0d0a432632dd755c3bea74ededea0619d +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..92fdc90c18815d9c22b41c21650399ae0343759f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d426ace1ca5421fc259c01bfd79180bbe5cdfc7bd4b796e00828010457442e33 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..94d6c5a589dd3c39bd39deb11b0e5e4f96dd736d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c83a37072b1808e469343f57559d31ce3929503dc3811a85a5a42679611c3b7c +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..b68205c575bd5490955dcfa1cd063fbb48a82438 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb27032a62c16f6a117bd2e85c89f4d1b272c3ac5b94a1352c5a6816c78ddc8f +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..262450393b4af9c8e06fe55890185cdfea10912f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b69ca38c0bb75456b1a48beda9c6d53ff29e394c82869433936bd6cc5e9193f +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2304/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f90b8c27b6972bbb4a1ca0bc0dffdc47237e677d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b057fe862d340f767a25cc1e8c8a0bb24146431f818f82ef2cc9b3d15cf7bfad +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2304/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..3a478f1679d0f2987a542859944e53bb753ddd9a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/trainer_state.json @@ -0,0 +1,16410 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 32.0, + "eval_steps": 500, + "global_step": 2304, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9.822762756263117e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2304/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2304/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2304/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2592/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2592/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b8932b2095bc39e6c64d178cd9d435f89a28ee9b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:159af0671f14991d93a0cf3de43bbb61f98764ddd20b8c4d113e6d53b75ea79e +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2592/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7b3be601378ad7adc03184c81d0c367d0fedfecf --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee5657c939446cb739c72fd10af9dfc17bf998cb81acec1a67f046885c0e8798 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..a0fdb39149ff58d0081c949f5c4ac75fad91bdea --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9fb29c856023c989256c2258f7638a00483c73b0e057199d87f49bde9adc001 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..a4a189614e11e6247011118d86c4300b64bec057 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8dcfc3024cdc624110f90d7ee29b0d17ef0cc4e4a51c6bb3887cfb1aa7d94b2b +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..3ca337805739ff7a42b8bda27c5bc07f46adbdf8 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f72e1db756ecb49e4b074f22485247cd548215c7452157ef8ff95dbe70728a87 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..387d4ca2cead233b29667cf1b24c44a90092d396 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e2fea0eba78e390537679e1702fa655f12d93c6e64ab30b2ec5f73f5825e46b +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2592/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f5637f631cc7ed2167f366c33ad859c0b7d97027 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0eee245b547fe91bdb4d6b4cbe52724d52118d751812b2781fcb358a63c34082 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2592/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..31e125ba45f1e965b57c37b69148d069d61b3643 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/trainer_state.json @@ -0,0 +1,18458 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 36.0, + "eval_steps": 500, + "global_step": 2592, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.1050608100796006e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2592/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2592/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2592/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-288/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..932e2392f1f7ca6b990b2f73c2370b3470207aff --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b73f8b5606f1126045d5c65fba98a80f8bf4d48c7ee9d040ff33f8e6df39d35 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..142e0688e153e7072df672f1970d3ffd4e3d39c9 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bf29c4366db961bc92ec5cbb00d9ecc6660ab160276331129e62e7dafe8bfce +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..1f77348e5e2510858cdf15d1d52ccae69d3c15ec --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4be5e023fa0cbf8e69b3989a76be3c51c1003595cef9a1cd6a981e96df33dc81 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..1be981878581dc4cc5e85e632fb1f24089818b49 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f07d0e39393ed810f745b19bf8ca8fa3347107f534dbdfb8793f199d1dba009c +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..6fbcd414b36d096739a953ed547b1dff430947f8 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48b06f9ecad190f56f234653cd52c124e3cf5e35fea91d2ea5b309bab6a46c3c +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..36c6f976b01de6004b9d1389ad29e9283efba9b3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed7c98726399543e24ee5ac4b4f3ea212ab29fcfec2a586bb2dc00d6f002e8de +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..e7f2e4bf5abc860ae45f2d5dbda8115542dd3611 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d1eab1d6f570939dc1e4b9d10383b7354ef244100fb2b34ff0ac8b58642b63c +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..534ab212ab85e7abb844693a88e2f7df98a1a732 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json @@ -0,0 +1,2074 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.0, + "eval_steps": 500, + "global_step": 288, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.2278453445328896e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-288/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-288/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-288/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-2880/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-2880/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7a51c5c1ccac4c8d7d276fc3783b893948605be4 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bb6f8b646626a86e79f911e3e834b2a10e372457b71dee43ebcd5f20db74bcf +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2880/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..c8cfb763c02c5f0dcc46e80ccc2e2afcaaa2c2ff --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2c25bdb61fa498e65d81733fbe7a6bc7da13085b041bd3dd615273bdbe1b41e +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..46aa5d638d90b38e6e23e319e89d3ad4c2f9f4d2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b9c0da4d03bbbcc4b70f3bfcb5cb2f9f49da5ef15eeea1bf8411056223551c1 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..c94fd8f927e40f13f0dddd0206934032e8478afc --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a2e93731721a79a7062c5c650729198de80400972629e91fe965623d5298d09 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..caabbc0e5bea10106f5ec02ac960e5854486d067 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ca72e3ce18e7d072c336c12a9ac59d6061211259635b7cea434b569953bcd7e +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..c24519070b860089a9d5b1e9e5f6ef121eb04590 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62b58282799b63651225abb40495daf3718399e4ae4f5d57549b5ba4ac2f8e58 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-2880/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0a5f79a713b34dd2d8d63f191629551c649dd599 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc3ec3427ea842499c98ae42d325c555883f86219e2410d0c1a47e90ade53b1c +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-2880/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..aebf5254a61e0691e8d39c066bb1d6807395e8df --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/trainer_state.json @@ -0,0 +1,20506 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 40.0, + "eval_steps": 500, + "global_step": 2880, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.2278453445328896e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-2880/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-2880/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-2880/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-3168/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-3168/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0dcfa3a30b5fa6752cd1e6f79057359c369bbfc6 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6e814bb9e83b141a3d571b77177fe21003a41a1673ed9f2cf38a18db57f9b1b +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3168/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5b2fa2f340cfdfa488848d3eaa140c5a4a1d51ee --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23dcb35da4d7c749c248d6d94637c0c19abadf9e3a4fd88122c9935ae5798d5d +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..5cead0762b939d4cfee78cbe3a60d54daa25efb0 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e593743986927a9cd1648c0e4bd6c7f8a21b611ad69fcae341973d2d7515960 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..77b5bcfd10229dee27d6ac33a93e7b8364000f70 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b03a37a26d716a393d503dc387470854bc62410e93ec96de951f0aff2c7d4cc +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..71f30120f7b4f413225d384ebd49b9888ae690a0 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dbf5cf5bc8065e48c4cdfbb7805e10a738b6f2e0271c2419220921c94cc9122 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..8f5ccdd68668ff317221c6f2f95b7735455ecd05 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e782ad710797b96d3abdb28995d2a41bd12fdcc553cfa86a98bfd43032b0b4b +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3168/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..197051978d6c74fdad6b7032be258e0e7820d474 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e78dacd009cacb02d7e6aed95e856956ae6e4311674c02b7eda2409a3c3c392 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-3168/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..23ae523aaa95147509a158856f3bec083db76ac5 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/trainer_state.json @@ -0,0 +1,22554 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 44.0, + "eval_steps": 500, + "global_step": 3168, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.3506298789861786e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3168/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-3168/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3168/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-3456/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-3456/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dc71bb4ad1af2d7517c8abc07c9a2e8720061ef9 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49654181507b5a62db885d2846536aee3275e38c88c34f4ebd1baf7d457f9ec6 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3456/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..11e11a544a685e991419998111d4098433f55c76 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01192f68bd9ec49948989234584a715cc3eb1e5cca3271bedf3d0144aa1e34a4 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..53ddb8e1764017752759402da1b95202b3baf1e9 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dcb6653f8376fc990944caeb4b9eb7aae01fb54a7598cdc206beb81b83cb2ff +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..cc84b3583e27ed3fb5357880884ee8b623a61531 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cd30828cea7b57651e4ff7a62b2441b5b0fe5521a3553e12bcdcf4d7da20691 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..7d556d66f69473a5f0c0dbde9bba5ed9b61a5c4f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:614c6993219e973b42f4e87c341515ecd80e83ed9ee205cbf61430d4c2357d3c +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..29050e842dcc4976e1dd2e1877e62d2227ea3c2f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2ca26b0a3f3294ddb7d6e30792aae3898215fe7ebe6d34ae1498c30ac0653ff +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3456/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6f299b34b59cabea3196797084c0dd37275dcaa8 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a30063d3a42c95cfa4d505a809336bc072a917980dcfbeb28708fcb4baba115 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-3456/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d01427abd2f15400b92d1e83ffb55f200a570544 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/trainer_state.json @@ -0,0 +1,24602 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 48.0, + "eval_steps": 500, + "global_step": 3456, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.4734144134394675e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3456/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-3456/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3456/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-3744/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-3744/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2ec81ab0377664c10aabba3af0bf3b69f93ae2cd --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bc2b51e2915002f43ce9df305c5014f83c702527028ecdf03c2c77dec88feee +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3744/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3ffbfe0662338dce32c7ee31990f1a1a1ea93d1c --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c7b20cb99576253b87386125af4bbb9ad1121eb8df6cea4e611b07ed54ea615 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..c097e0911dd435e7778913d466a4c8e8b90d9b84 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1176d244adaa134fc1c33fc3d5f23966bc402da63e8d71cc7436503fabd30d43 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..fb77759bb4236393d9b44cac5502638baa61429e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:752f121ae614cecac88411665432466278a4bb5b3fc46ca2ef8287271608db27 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..c4e6038a511b50d7a59a55c7ef57633938b53fd0 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:350de1630844c162a2e13816d52724a1a18ced3552347a5ca57068d1ee6c50b7 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..b675dc9296475644de03638efd8b8109d32ca1ef --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8923c51953991a359e189cde40248a1821c5e36a0f0dc6ee11bd455b08aba509 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-3744/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..20b086465f72e318302a946aa7b4256ef216a798 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7fdc869be493cb85f416c60210527a1c69abda3f8a295fcb1ff6f1bec552aedc +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-3744/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..3605246a0545a878f31c4eacd16e77ceb81c0800 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/trainer_state.json @@ -0,0 +1,26650 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 52.0, + "eval_steps": 500, + "global_step": 3744, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.5961989478927565e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-3744/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-3744/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-3744/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4032/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4032/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e100d59b7aaacd71434636a1c89885f21cc5fa50 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:412f049b6362640e7d8533ad087f506ac4b0493f32fa49f321584598b50f890d +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4032/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e80be47e4f07320bb653502a98217c84235dd68c --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93e3cc6b6dfeb78481954b5f343786b73d0665da20dfebd9ecc6a71153e60439 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..ad303665e91bd8f463c96f8ed5c5df78f614a9fa --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa32010f36bc8768606dd76bbf2cc108c36257b13d01a3f21156b47945e64631 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..4bac4005ebb4b7f2f74832f578c276bcb71f334a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a78e70ff4356ac6656e8073aba4a1df6fa153f4e3114fafaf49b3cc0a03475bf +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..0db8e27e712c3dd6632afa8e2d230a3da16e1364 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff498817004558cbfa2b8135a52cbe67932c55151d0c17d095105acdf1951053 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..951c60de2fb5f641f5a18648f8761a3af42de82f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:863bc081e419560c57a86100786cd6de3749e7265ca91548d5cd0f11e6d052b5 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4032/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..5da96a3852815ab990f23f4b91093acc83e76469 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9dd17660f6a4c270971d4c998ea51235588dc2ab05504036a1351d00b0a57d2c +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4032/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..dc148a0e461de1f5c1ee5b27a570890ef7358abe --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/trainer_state.json @@ -0,0 +1,28698 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 56.0, + "eval_steps": 500, + "global_step": 4032, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.7189834823460454e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4032/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4032/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4032/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4320/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4320/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81ef3fb2d97e60fe0e421a3ef81f0e91b4cb0c09 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d95648a6b9afbdcbd09ea254944ed25587d460ffb84366de3d076e2721ed01f2 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4320/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7d2d5d07f4e70ded0e545b44c4e704baad78f9d8 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bedc1e957284cd8f730f74e70592d488d054ec89f70926baad9378c399c3cb7 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..857a2f67a20363a76b04e398c4b43929a2876378 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b9c43b8c954569c82e76a9f2d993225b6b7dbf6b329599acfe5545ac32c8f22 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..d69ea8592d9d7a0ef2bbaa6e8f106b074b9c90f6 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26c963656cfadaeec7edb40d9c8f021310d2d6343fbbbe402907b79bfb5ada91 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..61a4d72e3ecceb047507833e7b22210803abc58e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f58ab0058bd189237a3838eb734c7827116ce23174e1b3a6a5c22a152fad5b3 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..2302ff29e4b6b17d0e81ad6016f44d9c82e028e9 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12cbf2c759fd06b8222f1c95276e256443f6c608d42df51778c153c282006dac +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4320/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..8304bf8d1ccccae077ecc7e6e66bba8073475ac5 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c504cbc081ac24362a030317097aa4f7d2e9075cfc12b05dc64f4ee58e8558c +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4320/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..0d1ae75e409b2a7f2fdecaeb6a16c540c9d2987b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/trainer_state.json @@ -0,0 +1,30746 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 60.0, + "eval_steps": 500, + "global_step": 4320, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.8417680167993344e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4320/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4320/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4320/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4608/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4608/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a4e8790d4cd6c8f19a0256a5e946ed4d1a94109c --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b64b7e8382655d170b13d1a6577e69b6f8470911a15d5110093e386bff274fab +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4608/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..bf6e229277fa7ecf83e54904a5d52a400fac0f4b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:257a71904ee506eea0c7c4afbf5510ffc96cd430a8f34d09c3552bb3aaa7c6ea +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..f91567317ab848c2dff078899eb662dd51945bae --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82bad169f67148d86bef2a656e092bfc17814ad929da77d19575a4e4d385d43c +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..ed61f380a617838b1ac7d8ff8c039ec4d4cbe5c6 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d116daf8954ab7a831f7098c4b5c2ad6a178ca829b12cf6ed99988af6804c5b +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..d68e42e69f3b5f81be555c974a3b3df909c2ea1d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d320b2fa71fd92a501be41c393f5297bdc9d88e31c0a36dfc68fb94a01290b14 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..ed2916c2a4e214438982b7fcbcb279f2806549e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56f75bbaa62722ad1b7a78bdf0b5790f29b39f7cbeaf0c2d6fc2de4d46c7b867 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4608/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..e577101181b24dfcc9b18d0a62144d06cb4b37c2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d97e4c65a4c8c2376a5585ec88bf61c1243526bb687aa9ffb5e9d7512df2daa +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4608/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..46f1e702d3ce871749b48bd1c1214a16fa5388a1 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/trainer_state.json @@ -0,0 +1,32794 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 64.0, + "eval_steps": 500, + "global_step": 4608, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.9645525512526234e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4608/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4608/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4608/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-4896/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-4896/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6eda745eca2b3b50d1ce4441fcc02cb99bf30786 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5df3108be16e0dcc8f841715dc42ed32fde017827e7e67080a1ea4d3447ca624 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4896/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..13d3c7cfe6712d55c65b7a4a71190481285bca97 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:159e7c5f95d1fd59d1a34afda01a4d02812101e4b25682aa81ddaa510d1e77ce +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..750ff4ede83419f049eeb660a694b7e3ecc7c8d6 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fa62fb1de6e4743a12d17318debd141345b5a642afd38db7758d0634c7a7f13 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..9ff82b1335a99fc6b3002849e9ff7716b929c804 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f69ce94754005caa11960e9c7c89ce4524851a59ae27cb35de69f711d23b218 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..088a0e343cacf37055aca354fb38e71ffd425531 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:993b3848c8a8140c118f377b7499a5492beb127bdfe31a0d905e551e1c466859 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..0610ab758a90414848478fab4b81e86746f261c3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bd27efae4d1ce0c2c890f064a5e19e1258acc7c189fd707628a2140a1a99d14 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-4896/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f446f829d82a1f86c912714abb53d7fba343010b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94642e081368d7b75b2ef306469e253305fe85f7bdeb714618031927a58d2a80 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-4896/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e7d06748e87af64234d992b80f80e04c74e9f768 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/trainer_state.json @@ -0,0 +1,34842 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 68.0, + "eval_steps": 500, + "global_step": 4896, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.0873370857059123e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-4896/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-4896/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-4896/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-5184/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-5184/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d6964b90321ecf947102216d4e9befa10b7992d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdd0baa43cae174cd7ab844b8496cdaad7bac65f887d1fd0c1428e0137c373f7 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5184/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..dc2838fe5fc58605fa80a338d826ad594ac2feae --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:679c528f3a002006294bb8deab8424c8ccf33503816d9605c2ae50708cb95ea4 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..8330111548d3ceb55eec2ce0d100fb6911e3f119 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22a799ebf18d5949b0d687c2fca394611640bd54f2edeab8c785f7bfd8ca1210 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..94c164b966d4b94a3d97959474303b95d0335ab3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b86e8ab98f075cb83d5d0c3f8da246a7181ec97b387d2d0b3b2596f3a4026908 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..f938712cd32e74263ab38e7e4e77dc3a52a35a56 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7305ed0dd577328473d6a79b96b2c5846f83995e5a3b25f37615248aa85284b5 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..b29fdbf1cbfdea70acf84092a9920a959ea9ee42 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f01527c462cf991ac25914ec71ab83c5feaff9e269428a30e5c3f3dcc48b907e +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5184/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..87c25f34fb9c3e31517de12eec62aa55b15b7f35 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f27f1d0b18f28ca972b7026932ac402d35af3b241ee72040352067f9ee1e6988 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-5184/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..12afb136dca88a2b936bd3342d80e61693cebeb9 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/trainer_state.json @@ -0,0 +1,36890 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 72.0, + "eval_steps": 500, + "global_step": 5184, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.2101216201592013e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5184/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-5184/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5184/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-5472/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-5472/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c6b0ceeffaee9156184bcddec1e3318b24ac0b57 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d6f0df4c6768e67e40e8b29861a944d8646a122543c6f0481e8117c070a0767 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5472/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..c8babd07599f242150deac02aa8cad8e803e7b3d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43ff9f559aba664ab73241955a5fb27153ab1285ca17c90bc5e46c614a4ad698 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..aff14f8ce1841e93fa324277ef2fcf672ce6f320 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07649dfce15de07bfbfc987960ffc4fc589f7b76aa2b77ac1b4369c58a4f5a9f +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..e5641e294b0c012dcee9b441e5a618d691fac622 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84c1f1ab67a99075a008b840c7088c557b3bd74f7085e4eb57ae5933f97e3096 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..a861486932b28d15d049b3b3b1850788353f236e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c246352697b88657a8da26487f5ced4f002f630719de68950b06addd63ed20c3 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..6165592a8f1c0944664783a010f42e4ab136086b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4cf00673b1da661d86e94245cd02b21123f7422892904494df18097422747aa +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5472/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..bde6e47e59bea2a2d1c7be49dbaa3a2d4dfadf37 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec3fa8fa903a5a893f5b74cfd95272dfde1a7a2149fd0f3ecec1adb7304a55b8 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-5472/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5887924a875c518f229ecdaa38b4d886282f279f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/trainer_state.json @@ -0,0 +1,38938 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 76.0, + "eval_steps": 500, + "global_step": 5472, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.33290615461249e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5472/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-5472/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5472/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-576/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4024774d232f7817ed4648e017722ba7708e5810 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5f0056e676bbae080ef7dddaf146e32ce36d296fe6594a7327e78b97be583a2 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..65065a006c00793f3c83b355a9dff47bb804851b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ed71010198015ed1b30f0465125e93c395a3cfe700f8c774df921c6b6f48403 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..112b041e1d36ec1494d292f68d4ed5a663a7760c --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16f4eb7f2a605fbab21186f755abda641fe39281304fff9e3cd8d1f68ad970b0 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..5b7627425b53209442125436df436e61e1a84ab1 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3ecca285a09feeb269b66fce026683d05f7aded4f691e94faec2607d18d85d8 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..bef918e95a7666f775e2c4b5e8379f56bf2b1155 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c164f926b8a1118b5abd6db76ed1ce8a68afbda1403f138f3634fa1a937f448d +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..1277d7e0e0037e1991a7937de0c59863261bf553 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b4476d524aa476d8f8153629400bf093c94260913a40a182ea50a014b63af36 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..dbc454a52c5d1edb241846f7cbf86268ced2c2ff --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f28b8022dab4a12fbbad2bccde532c0f7c60e469bde0bf66d40f7d043ac1727 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b176ef27d117d084ce630325cf416e0d647090c2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json @@ -0,0 +1,4122 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 8.0, + "eval_steps": 500, + "global_step": 576, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.4556906890657792e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-576/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-576/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-576/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-5760/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-5760/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c32862e1ebfabd231188cab38c0f5fa4ff1aeda --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44c96e9d16f833e2f029812687786b0c582ce8043d4854729e759d91e4ae5379 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5760/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..eca2cb0191673025aff2d7629b3647f70531dc3f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8927b1690f33447659d31c9807fb5f51447a30aba0302a83a57d0e44d2673dc8 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..a285714799f193b1956b942912d36ac13f0bf7ce --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d4b2db3d31ab56483f8ea333d381b8d330a869c9278fe352f6cbfd94741b766 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..2e2e42c1aab8de141eefa799062d967e975b6218 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1322bae719fba49418ac08f03776405c298a53bdd11d4473f4a6a187dc25005e +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..f2cefc91b060c9f1a03313ee35052d85ae1fb721 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f9d34a9ab58ec8211a1996d0791f9c8c7a6150afc12696230c213c6c98fd63b +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..8ad2dc824fec084bf990db595b420b6fcc0aff87 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73b5916c48044cca60d11397b7319bdd569e58f962208e54da1468f10cf9a5f4 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-5760/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..146fac8459db1a89ef1d0b02179a40f96280dd47 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a202cad58a75ceda9326b04df06bcdb148f3de25b706c12232648bdbcae46e67 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-5760/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..493d844116fb14245dcf82d371caeffd003511cd --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/trainer_state.json @@ -0,0 +1,40986 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 80.0, + "eval_steps": 500, + "global_step": 5760, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + }, + { + "epoch": 76.0, + "eval_loss": 4.689640998840332, + "eval_runtime": 56.9095, + "eval_samples_per_second": 42.91, + "eval_steps_per_second": 1.353, + "step": 5472 + }, + { + "epoch": 76.01397379912663, + "grad_norm": 0.015351055189967155, + "learning_rate": 0.0006, + "loss": 4.348268985748291, + "step": 5473 + }, + { + "epoch": 76.02794759825328, + "grad_norm": 0.017164217308163643, + "learning_rate": 0.0006, + "loss": 4.244204998016357, + "step": 5474 + }, + { + "epoch": 76.04192139737991, + "grad_norm": 0.021624932065606117, + "learning_rate": 0.0006, + "loss": 4.311168670654297, + "step": 5475 + }, + { + "epoch": 76.05589519650655, + "grad_norm": 0.028712468221783638, + "learning_rate": 0.0006, + "loss": 4.322287082672119, + "step": 5476 + }, + { + "epoch": 76.06986899563319, + "grad_norm": 0.027915487065911293, + "learning_rate": 0.0006, + "loss": 4.381207466125488, + "step": 5477 + }, + { + "epoch": 76.08384279475983, + "grad_norm": 0.017955303192138672, + "learning_rate": 0.0006, + "loss": 4.2635016441345215, + "step": 5478 + }, + { + "epoch": 76.09781659388646, + "grad_norm": 0.019113514572381973, + "learning_rate": 0.0006, + "loss": 4.399471282958984, + "step": 5479 + }, + { + "epoch": 76.1117903930131, + "grad_norm": 0.017502542585134506, + "learning_rate": 0.0006, + "loss": 4.321204662322998, + "step": 5480 + }, + { + "epoch": 76.12576419213974, + "grad_norm": 0.01852525770664215, + "learning_rate": 0.0006, + "loss": 4.217884063720703, + "step": 5481 + }, + { + "epoch": 76.13973799126637, + "grad_norm": 0.019836828112602234, + "learning_rate": 0.0006, + "loss": 4.278532028198242, + "step": 5482 + }, + { + "epoch": 76.15371179039302, + "grad_norm": 0.02101975679397583, + "learning_rate": 0.0006, + "loss": 4.246720314025879, + "step": 5483 + }, + { + "epoch": 76.16768558951965, + "grad_norm": 0.021202990785241127, + "learning_rate": 0.0006, + "loss": 4.517517566680908, + "step": 5484 + }, + { + "epoch": 76.18165938864628, + "grad_norm": 0.021428769454360008, + "learning_rate": 0.0006, + "loss": 4.3373517990112305, + "step": 5485 + }, + { + "epoch": 76.19563318777293, + "grad_norm": 0.018263190984725952, + "learning_rate": 0.0006, + "loss": 4.402856826782227, + "step": 5486 + }, + { + "epoch": 76.20960698689956, + "grad_norm": 0.019025536254048347, + "learning_rate": 0.0006, + "loss": 4.314227104187012, + "step": 5487 + }, + { + "epoch": 76.2235807860262, + "grad_norm": 0.020918630063533783, + "learning_rate": 0.0006, + "loss": 4.454369068145752, + "step": 5488 + }, + { + "epoch": 76.23755458515284, + "grad_norm": 0.019307617098093033, + "learning_rate": 0.0006, + "loss": 4.247509002685547, + "step": 5489 + }, + { + "epoch": 76.25152838427948, + "grad_norm": 0.014956331811845303, + "learning_rate": 0.0006, + "loss": 4.365175247192383, + "step": 5490 + }, + { + "epoch": 76.26550218340611, + "grad_norm": 0.017213059589266777, + "learning_rate": 0.0006, + "loss": 4.384005069732666, + "step": 5491 + }, + { + "epoch": 76.27947598253274, + "grad_norm": 0.017575597390532494, + "learning_rate": 0.0006, + "loss": 4.41934871673584, + "step": 5492 + }, + { + "epoch": 76.29344978165939, + "grad_norm": 0.015879256650805473, + "learning_rate": 0.0006, + "loss": 4.26675271987915, + "step": 5493 + }, + { + "epoch": 76.30742358078602, + "grad_norm": 0.01564522460103035, + "learning_rate": 0.0006, + "loss": 4.277339935302734, + "step": 5494 + }, + { + "epoch": 76.32139737991267, + "grad_norm": 0.016414586454629898, + "learning_rate": 0.0006, + "loss": 4.4138593673706055, + "step": 5495 + }, + { + "epoch": 76.3353711790393, + "grad_norm": 0.014732337556779385, + "learning_rate": 0.0006, + "loss": 4.281006336212158, + "step": 5496 + }, + { + "epoch": 76.34934497816595, + "grad_norm": 0.015461455099284649, + "learning_rate": 0.0006, + "loss": 4.263422966003418, + "step": 5497 + }, + { + "epoch": 76.36331877729258, + "grad_norm": 0.014529814012348652, + "learning_rate": 0.0006, + "loss": 4.36165189743042, + "step": 5498 + }, + { + "epoch": 76.37729257641921, + "grad_norm": 0.01469483319669962, + "learning_rate": 0.0006, + "loss": 4.309642791748047, + "step": 5499 + }, + { + "epoch": 76.39126637554585, + "grad_norm": 0.016490574926137924, + "learning_rate": 0.0006, + "loss": 4.288150787353516, + "step": 5500 + }, + { + "epoch": 76.40524017467249, + "grad_norm": 0.014593565836548805, + "learning_rate": 0.0006, + "loss": 4.271614074707031, + "step": 5501 + }, + { + "epoch": 76.41921397379913, + "grad_norm": 0.012796535156667233, + "learning_rate": 0.0006, + "loss": 4.420042037963867, + "step": 5502 + }, + { + "epoch": 76.43318777292576, + "grad_norm": 0.016086025163531303, + "learning_rate": 0.0006, + "loss": 4.289943695068359, + "step": 5503 + }, + { + "epoch": 76.44716157205241, + "grad_norm": 0.015439470298588276, + "learning_rate": 0.0006, + "loss": 4.3023271560668945, + "step": 5504 + }, + { + "epoch": 76.46113537117904, + "grad_norm": 0.015066367574036121, + "learning_rate": 0.0006, + "loss": 4.197081565856934, + "step": 5505 + }, + { + "epoch": 76.47510917030567, + "grad_norm": 0.016653263941407204, + "learning_rate": 0.0006, + "loss": 4.333785057067871, + "step": 5506 + }, + { + "epoch": 76.48908296943232, + "grad_norm": 0.014791973866522312, + "learning_rate": 0.0006, + "loss": 4.323811054229736, + "step": 5507 + }, + { + "epoch": 76.50305676855895, + "grad_norm": 0.014694827608764172, + "learning_rate": 0.0006, + "loss": 4.195852279663086, + "step": 5508 + }, + { + "epoch": 76.5170305676856, + "grad_norm": 0.015120396390557289, + "learning_rate": 0.0006, + "loss": 4.42379903793335, + "step": 5509 + }, + { + "epoch": 76.53100436681223, + "grad_norm": 0.014584025368094444, + "learning_rate": 0.0006, + "loss": 4.2702531814575195, + "step": 5510 + }, + { + "epoch": 76.54497816593886, + "grad_norm": 0.013981613330543041, + "learning_rate": 0.0006, + "loss": 4.383029937744141, + "step": 5511 + }, + { + "epoch": 76.5589519650655, + "grad_norm": 0.012930642813444138, + "learning_rate": 0.0006, + "loss": 4.296136856079102, + "step": 5512 + }, + { + "epoch": 76.57292576419214, + "grad_norm": 0.013060989789664745, + "learning_rate": 0.0006, + "loss": 4.346380233764648, + "step": 5513 + }, + { + "epoch": 76.58689956331878, + "grad_norm": 0.013209798373281956, + "learning_rate": 0.0006, + "loss": 4.316067218780518, + "step": 5514 + }, + { + "epoch": 76.60087336244541, + "grad_norm": 0.012423216365277767, + "learning_rate": 0.0006, + "loss": 4.370274543762207, + "step": 5515 + }, + { + "epoch": 76.61484716157206, + "grad_norm": 0.013205411843955517, + "learning_rate": 0.0006, + "loss": 4.429969310760498, + "step": 5516 + }, + { + "epoch": 76.62882096069869, + "grad_norm": 0.013049827888607979, + "learning_rate": 0.0006, + "loss": 4.378549098968506, + "step": 5517 + }, + { + "epoch": 76.64279475982532, + "grad_norm": 0.012075880542397499, + "learning_rate": 0.0006, + "loss": 4.348293304443359, + "step": 5518 + }, + { + "epoch": 76.65676855895197, + "grad_norm": 0.014932847581803799, + "learning_rate": 0.0006, + "loss": 4.282538890838623, + "step": 5519 + }, + { + "epoch": 76.6707423580786, + "grad_norm": 0.015615998767316341, + "learning_rate": 0.0006, + "loss": 4.274641036987305, + "step": 5520 + }, + { + "epoch": 76.68471615720524, + "grad_norm": 0.01424513477832079, + "learning_rate": 0.0006, + "loss": 4.243166923522949, + "step": 5521 + }, + { + "epoch": 76.69868995633188, + "grad_norm": 0.012775546871125698, + "learning_rate": 0.0006, + "loss": 4.335611820220947, + "step": 5522 + }, + { + "epoch": 76.71266375545852, + "grad_norm": 0.014461303129792213, + "learning_rate": 0.0006, + "loss": 4.246976375579834, + "step": 5523 + }, + { + "epoch": 76.72663755458515, + "grad_norm": 0.0145436255261302, + "learning_rate": 0.0006, + "loss": 4.275895118713379, + "step": 5524 + }, + { + "epoch": 76.74061135371178, + "grad_norm": 0.014055085368454456, + "learning_rate": 0.0006, + "loss": 4.331108093261719, + "step": 5525 + }, + { + "epoch": 76.75458515283843, + "grad_norm": 0.011933894827961922, + "learning_rate": 0.0006, + "loss": 4.293721675872803, + "step": 5526 + }, + { + "epoch": 76.76855895196506, + "grad_norm": 0.01277392078191042, + "learning_rate": 0.0006, + "loss": 4.312374114990234, + "step": 5527 + }, + { + "epoch": 76.78253275109171, + "grad_norm": 0.014227936044335365, + "learning_rate": 0.0006, + "loss": 4.238729000091553, + "step": 5528 + }, + { + "epoch": 76.79650655021834, + "grad_norm": 0.01488024927675724, + "learning_rate": 0.0006, + "loss": 4.24587345123291, + "step": 5529 + }, + { + "epoch": 76.81048034934499, + "grad_norm": 0.015341250225901604, + "learning_rate": 0.0006, + "loss": 4.257378578186035, + "step": 5530 + }, + { + "epoch": 76.82445414847162, + "grad_norm": 0.015179437585175037, + "learning_rate": 0.0006, + "loss": 4.2812910079956055, + "step": 5531 + }, + { + "epoch": 76.83842794759825, + "grad_norm": 0.012814724817872047, + "learning_rate": 0.0006, + "loss": 4.2309465408325195, + "step": 5532 + }, + { + "epoch": 76.8524017467249, + "grad_norm": 0.013810716569423676, + "learning_rate": 0.0006, + "loss": 4.325292110443115, + "step": 5533 + }, + { + "epoch": 76.86637554585153, + "grad_norm": 0.017468232661485672, + "learning_rate": 0.0006, + "loss": 4.243788719177246, + "step": 5534 + }, + { + "epoch": 76.88034934497817, + "grad_norm": 0.01739857904613018, + "learning_rate": 0.0006, + "loss": 4.345390796661377, + "step": 5535 + }, + { + "epoch": 76.8943231441048, + "grad_norm": 0.015329061076045036, + "learning_rate": 0.0006, + "loss": 4.415142059326172, + "step": 5536 + }, + { + "epoch": 76.90829694323143, + "grad_norm": 0.01598426327109337, + "learning_rate": 0.0006, + "loss": 4.299558639526367, + "step": 5537 + }, + { + "epoch": 76.92227074235808, + "grad_norm": 0.01621454581618309, + "learning_rate": 0.0006, + "loss": 4.284354209899902, + "step": 5538 + }, + { + "epoch": 76.93624454148471, + "grad_norm": 0.015956968069076538, + "learning_rate": 0.0006, + "loss": 4.293837547302246, + "step": 5539 + }, + { + "epoch": 76.95021834061136, + "grad_norm": 0.01776493526995182, + "learning_rate": 0.0006, + "loss": 4.229729175567627, + "step": 5540 + }, + { + "epoch": 76.96419213973799, + "grad_norm": 0.015317477285861969, + "learning_rate": 0.0006, + "loss": 4.3328447341918945, + "step": 5541 + }, + { + "epoch": 76.97816593886463, + "grad_norm": 0.015140251256525517, + "learning_rate": 0.0006, + "loss": 4.310840606689453, + "step": 5542 + }, + { + "epoch": 76.99213973799127, + "grad_norm": 0.015326812863349915, + "learning_rate": 0.0006, + "loss": 4.230832099914551, + "step": 5543 + }, + { + "epoch": 77.0, + "grad_norm": 0.01607763208448887, + "learning_rate": 0.0006, + "loss": 4.2236328125, + "step": 5544 + }, + { + "epoch": 77.0, + "eval_loss": 4.662221431732178, + "eval_runtime": 56.5194, + "eval_samples_per_second": 43.206, + "eval_steps_per_second": 1.362, + "step": 5544 + }, + { + "epoch": 77.01397379912663, + "grad_norm": 0.014839510433375835, + "learning_rate": 0.0006, + "loss": 4.254111289978027, + "step": 5545 + }, + { + "epoch": 77.02794759825328, + "grad_norm": 0.01519741676747799, + "learning_rate": 0.0006, + "loss": 4.1607184410095215, + "step": 5546 + }, + { + "epoch": 77.04192139737991, + "grad_norm": 0.015472603030502796, + "learning_rate": 0.0006, + "loss": 4.323320388793945, + "step": 5547 + }, + { + "epoch": 77.05589519650655, + "grad_norm": 0.017937948927283287, + "learning_rate": 0.0006, + "loss": 4.308490753173828, + "step": 5548 + }, + { + "epoch": 77.06986899563319, + "grad_norm": 0.020890841260552406, + "learning_rate": 0.0006, + "loss": 4.259101867675781, + "step": 5549 + }, + { + "epoch": 77.08384279475983, + "grad_norm": 0.01892857998609543, + "learning_rate": 0.0006, + "loss": 4.386970520019531, + "step": 5550 + }, + { + "epoch": 77.09781659388646, + "grad_norm": 0.016583630815148354, + "learning_rate": 0.0006, + "loss": 4.233604907989502, + "step": 5551 + }, + { + "epoch": 77.1117903930131, + "grad_norm": 0.016063468530774117, + "learning_rate": 0.0006, + "loss": 4.339544296264648, + "step": 5552 + }, + { + "epoch": 77.12576419213974, + "grad_norm": 0.018326779827475548, + "learning_rate": 0.0006, + "loss": 4.367204666137695, + "step": 5553 + }, + { + "epoch": 77.13973799126637, + "grad_norm": 0.021590933203697205, + "learning_rate": 0.0006, + "loss": 4.352685928344727, + "step": 5554 + }, + { + "epoch": 77.15371179039302, + "grad_norm": 0.019643642008304596, + "learning_rate": 0.0006, + "loss": 4.258535385131836, + "step": 5555 + }, + { + "epoch": 77.16768558951965, + "grad_norm": 0.018476149067282677, + "learning_rate": 0.0006, + "loss": 4.232568740844727, + "step": 5556 + }, + { + "epoch": 77.18165938864628, + "grad_norm": 0.01584405079483986, + "learning_rate": 0.0006, + "loss": 4.258371353149414, + "step": 5557 + }, + { + "epoch": 77.19563318777293, + "grad_norm": 0.01575201191008091, + "learning_rate": 0.0006, + "loss": 4.183218002319336, + "step": 5558 + }, + { + "epoch": 77.20960698689956, + "grad_norm": 0.016871824860572815, + "learning_rate": 0.0006, + "loss": 4.194563865661621, + "step": 5559 + }, + { + "epoch": 77.2235807860262, + "grad_norm": 0.02059333398938179, + "learning_rate": 0.0006, + "loss": 4.195281982421875, + "step": 5560 + }, + { + "epoch": 77.23755458515284, + "grad_norm": 0.022640453651547432, + "learning_rate": 0.0006, + "loss": 4.279615879058838, + "step": 5561 + }, + { + "epoch": 77.25152838427948, + "grad_norm": 0.020335959270596504, + "learning_rate": 0.0006, + "loss": 4.25040864944458, + "step": 5562 + }, + { + "epoch": 77.26550218340611, + "grad_norm": 0.020010793581604958, + "learning_rate": 0.0006, + "loss": 4.291116714477539, + "step": 5563 + }, + { + "epoch": 77.27947598253274, + "grad_norm": 0.02109045907855034, + "learning_rate": 0.0006, + "loss": 4.364088535308838, + "step": 5564 + }, + { + "epoch": 77.29344978165939, + "grad_norm": 0.019519446417689323, + "learning_rate": 0.0006, + "loss": 4.329166412353516, + "step": 5565 + }, + { + "epoch": 77.30742358078602, + "grad_norm": 0.01689804159104824, + "learning_rate": 0.0006, + "loss": 4.244176864624023, + "step": 5566 + }, + { + "epoch": 77.32139737991267, + "grad_norm": 0.01893681474030018, + "learning_rate": 0.0006, + "loss": 4.239068984985352, + "step": 5567 + }, + { + "epoch": 77.3353711790393, + "grad_norm": 0.015998193994164467, + "learning_rate": 0.0006, + "loss": 4.173398017883301, + "step": 5568 + }, + { + "epoch": 77.34934497816595, + "grad_norm": 0.013892280869185925, + "learning_rate": 0.0006, + "loss": 4.287654876708984, + "step": 5569 + }, + { + "epoch": 77.36331877729258, + "grad_norm": 0.017060643061995506, + "learning_rate": 0.0006, + "loss": 4.398447036743164, + "step": 5570 + }, + { + "epoch": 77.37729257641921, + "grad_norm": 0.015168849378824234, + "learning_rate": 0.0006, + "loss": 4.15665864944458, + "step": 5571 + }, + { + "epoch": 77.39126637554585, + "grad_norm": 0.014886523596942425, + "learning_rate": 0.0006, + "loss": 4.196673393249512, + "step": 5572 + }, + { + "epoch": 77.40524017467249, + "grad_norm": 0.016941796988248825, + "learning_rate": 0.0006, + "loss": 4.314562797546387, + "step": 5573 + }, + { + "epoch": 77.41921397379913, + "grad_norm": 0.015353812836110592, + "learning_rate": 0.0006, + "loss": 4.409579277038574, + "step": 5574 + }, + { + "epoch": 77.43318777292576, + "grad_norm": 0.015192852355539799, + "learning_rate": 0.0006, + "loss": 4.392045974731445, + "step": 5575 + }, + { + "epoch": 77.44716157205241, + "grad_norm": 0.014232569374144077, + "learning_rate": 0.0006, + "loss": 4.21639347076416, + "step": 5576 + }, + { + "epoch": 77.46113537117904, + "grad_norm": 0.014453226700425148, + "learning_rate": 0.0006, + "loss": 4.2802276611328125, + "step": 5577 + }, + { + "epoch": 77.47510917030567, + "grad_norm": 0.012881236150860786, + "learning_rate": 0.0006, + "loss": 4.375980854034424, + "step": 5578 + }, + { + "epoch": 77.48908296943232, + "grad_norm": 0.01583976298570633, + "learning_rate": 0.0006, + "loss": 4.264520645141602, + "step": 5579 + }, + { + "epoch": 77.50305676855895, + "grad_norm": 0.014171584509313107, + "learning_rate": 0.0006, + "loss": 4.225279808044434, + "step": 5580 + }, + { + "epoch": 77.5170305676856, + "grad_norm": 0.014715159311890602, + "learning_rate": 0.0006, + "loss": 4.369834899902344, + "step": 5581 + }, + { + "epoch": 77.53100436681223, + "grad_norm": 0.016554079949855804, + "learning_rate": 0.0006, + "loss": 4.2976179122924805, + "step": 5582 + }, + { + "epoch": 77.54497816593886, + "grad_norm": 0.015095747075974941, + "learning_rate": 0.0006, + "loss": 4.294407844543457, + "step": 5583 + }, + { + "epoch": 77.5589519650655, + "grad_norm": 0.012726361863315105, + "learning_rate": 0.0006, + "loss": 4.221401214599609, + "step": 5584 + }, + { + "epoch": 77.57292576419214, + "grad_norm": 0.014999309554696083, + "learning_rate": 0.0006, + "loss": 4.243929386138916, + "step": 5585 + }, + { + "epoch": 77.58689956331878, + "grad_norm": 0.016526037827134132, + "learning_rate": 0.0006, + "loss": 4.374139785766602, + "step": 5586 + }, + { + "epoch": 77.60087336244541, + "grad_norm": 0.01648339442908764, + "learning_rate": 0.0006, + "loss": 4.392425060272217, + "step": 5587 + }, + { + "epoch": 77.61484716157206, + "grad_norm": 0.015269213356077671, + "learning_rate": 0.0006, + "loss": 4.33986759185791, + "step": 5588 + }, + { + "epoch": 77.62882096069869, + "grad_norm": 0.01690507121384144, + "learning_rate": 0.0006, + "loss": 4.307476043701172, + "step": 5589 + }, + { + "epoch": 77.64279475982532, + "grad_norm": 0.016438685357570648, + "learning_rate": 0.0006, + "loss": 4.2406206130981445, + "step": 5590 + }, + { + "epoch": 77.65676855895197, + "grad_norm": 0.0157600287348032, + "learning_rate": 0.0006, + "loss": 4.314070224761963, + "step": 5591 + }, + { + "epoch": 77.6707423580786, + "grad_norm": 0.015127947553992271, + "learning_rate": 0.0006, + "loss": 4.22475528717041, + "step": 5592 + }, + { + "epoch": 77.68471615720524, + "grad_norm": 0.016632677987217903, + "learning_rate": 0.0006, + "loss": 4.26954984664917, + "step": 5593 + }, + { + "epoch": 77.69868995633188, + "grad_norm": 0.013150460086762905, + "learning_rate": 0.0006, + "loss": 4.369131088256836, + "step": 5594 + }, + { + "epoch": 77.71266375545852, + "grad_norm": 0.014513005502521992, + "learning_rate": 0.0006, + "loss": 4.3248796463012695, + "step": 5595 + }, + { + "epoch": 77.72663755458515, + "grad_norm": 0.014090991578996181, + "learning_rate": 0.0006, + "loss": 4.268381118774414, + "step": 5596 + }, + { + "epoch": 77.74061135371178, + "grad_norm": 0.01633097417652607, + "learning_rate": 0.0006, + "loss": 4.238439083099365, + "step": 5597 + }, + { + "epoch": 77.75458515283843, + "grad_norm": 0.01710933819413185, + "learning_rate": 0.0006, + "loss": 4.399697780609131, + "step": 5598 + }, + { + "epoch": 77.76855895196506, + "grad_norm": 0.019251767545938492, + "learning_rate": 0.0006, + "loss": 4.290450096130371, + "step": 5599 + }, + { + "epoch": 77.78253275109171, + "grad_norm": 0.01849171705543995, + "learning_rate": 0.0006, + "loss": 4.267734527587891, + "step": 5600 + }, + { + "epoch": 77.79650655021834, + "grad_norm": 0.014173115603625774, + "learning_rate": 0.0006, + "loss": 4.288799285888672, + "step": 5601 + }, + { + "epoch": 77.81048034934499, + "grad_norm": 0.014572718180716038, + "learning_rate": 0.0006, + "loss": 4.3790435791015625, + "step": 5602 + }, + { + "epoch": 77.82445414847162, + "grad_norm": 0.016375336796045303, + "learning_rate": 0.0006, + "loss": 4.231725692749023, + "step": 5603 + }, + { + "epoch": 77.83842794759825, + "grad_norm": 0.017326701432466507, + "learning_rate": 0.0006, + "loss": 4.211495399475098, + "step": 5604 + }, + { + "epoch": 77.8524017467249, + "grad_norm": 0.016399390995502472, + "learning_rate": 0.0006, + "loss": 4.386699199676514, + "step": 5605 + }, + { + "epoch": 77.86637554585153, + "grad_norm": 0.018269989639520645, + "learning_rate": 0.0006, + "loss": 4.254856109619141, + "step": 5606 + }, + { + "epoch": 77.88034934497817, + "grad_norm": 0.018443435430526733, + "learning_rate": 0.0006, + "loss": 4.391961097717285, + "step": 5607 + }, + { + "epoch": 77.8943231441048, + "grad_norm": 0.01784636080265045, + "learning_rate": 0.0006, + "loss": 4.273881912231445, + "step": 5608 + }, + { + "epoch": 77.90829694323143, + "grad_norm": 0.01782994158565998, + "learning_rate": 0.0006, + "loss": 4.26833963394165, + "step": 5609 + }, + { + "epoch": 77.92227074235808, + "grad_norm": 0.016260741278529167, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 5610 + }, + { + "epoch": 77.93624454148471, + "grad_norm": 0.015482903458178043, + "learning_rate": 0.0006, + "loss": 4.187826633453369, + "step": 5611 + }, + { + "epoch": 77.95021834061136, + "grad_norm": 0.015193257480859756, + "learning_rate": 0.0006, + "loss": 4.286060333251953, + "step": 5612 + }, + { + "epoch": 77.96419213973799, + "grad_norm": 0.015110115520656109, + "learning_rate": 0.0006, + "loss": 4.2828192710876465, + "step": 5613 + }, + { + "epoch": 77.97816593886463, + "grad_norm": 0.015973486006259918, + "learning_rate": 0.0006, + "loss": 4.317939758300781, + "step": 5614 + }, + { + "epoch": 77.99213973799127, + "grad_norm": 0.017363594844937325, + "learning_rate": 0.0006, + "loss": 4.287721633911133, + "step": 5615 + }, + { + "epoch": 78.0, + "grad_norm": 0.015277642756700516, + "learning_rate": 0.0006, + "loss": 4.356618881225586, + "step": 5616 + }, + { + "epoch": 78.0, + "eval_loss": 4.642678737640381, + "eval_runtime": 56.7335, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 5616 + }, + { + "epoch": 78.01397379912663, + "grad_norm": 0.016794051975011826, + "learning_rate": 0.0006, + "loss": 4.279930591583252, + "step": 5617 + }, + { + "epoch": 78.02794759825328, + "grad_norm": 0.01632673107087612, + "learning_rate": 0.0006, + "loss": 4.235943794250488, + "step": 5618 + }, + { + "epoch": 78.04192139737991, + "grad_norm": 0.01437876746058464, + "learning_rate": 0.0006, + "loss": 4.2542243003845215, + "step": 5619 + }, + { + "epoch": 78.05589519650655, + "grad_norm": 0.014993119053542614, + "learning_rate": 0.0006, + "loss": 4.262046813964844, + "step": 5620 + }, + { + "epoch": 78.06986899563319, + "grad_norm": 0.015721119940280914, + "learning_rate": 0.0006, + "loss": 4.173436164855957, + "step": 5621 + }, + { + "epoch": 78.08384279475983, + "grad_norm": 0.01334085687994957, + "learning_rate": 0.0006, + "loss": 4.322175979614258, + "step": 5622 + }, + { + "epoch": 78.09781659388646, + "grad_norm": 0.015561388805508614, + "learning_rate": 0.0006, + "loss": 4.242125988006592, + "step": 5623 + }, + { + "epoch": 78.1117903930131, + "grad_norm": 0.014418602921068668, + "learning_rate": 0.0006, + "loss": 4.266186714172363, + "step": 5624 + }, + { + "epoch": 78.12576419213974, + "grad_norm": 0.013739909045398235, + "learning_rate": 0.0006, + "loss": 4.231633186340332, + "step": 5625 + }, + { + "epoch": 78.13973799126637, + "grad_norm": 0.01402067206799984, + "learning_rate": 0.0006, + "loss": 4.188406944274902, + "step": 5626 + }, + { + "epoch": 78.15371179039302, + "grad_norm": 0.013639481738209724, + "learning_rate": 0.0006, + "loss": 4.1586594581604, + "step": 5627 + }, + { + "epoch": 78.16768558951965, + "grad_norm": 0.012858159840106964, + "learning_rate": 0.0006, + "loss": 4.360448360443115, + "step": 5628 + }, + { + "epoch": 78.18165938864628, + "grad_norm": 0.013089739717543125, + "learning_rate": 0.0006, + "loss": 4.344371795654297, + "step": 5629 + }, + { + "epoch": 78.19563318777293, + "grad_norm": 0.01382314320653677, + "learning_rate": 0.0006, + "loss": 4.20378303527832, + "step": 5630 + }, + { + "epoch": 78.20960698689956, + "grad_norm": 0.014206619001924992, + "learning_rate": 0.0006, + "loss": 4.239891529083252, + "step": 5631 + }, + { + "epoch": 78.2235807860262, + "grad_norm": 0.01472168043255806, + "learning_rate": 0.0006, + "loss": 4.282382488250732, + "step": 5632 + }, + { + "epoch": 78.23755458515284, + "grad_norm": 0.014237025752663612, + "learning_rate": 0.0006, + "loss": 4.241545677185059, + "step": 5633 + }, + { + "epoch": 78.25152838427948, + "grad_norm": 0.013231472112238407, + "learning_rate": 0.0006, + "loss": 4.184896469116211, + "step": 5634 + }, + { + "epoch": 78.26550218340611, + "grad_norm": 0.013538471423089504, + "learning_rate": 0.0006, + "loss": 4.325181007385254, + "step": 5635 + }, + { + "epoch": 78.27947598253274, + "grad_norm": 0.013541961088776588, + "learning_rate": 0.0006, + "loss": 4.368722915649414, + "step": 5636 + }, + { + "epoch": 78.29344978165939, + "grad_norm": 0.013446723110973835, + "learning_rate": 0.0006, + "loss": 4.301422595977783, + "step": 5637 + }, + { + "epoch": 78.30742358078602, + "grad_norm": 0.014753809198737144, + "learning_rate": 0.0006, + "loss": 4.259127616882324, + "step": 5638 + }, + { + "epoch": 78.32139737991267, + "grad_norm": 0.013256008736789227, + "learning_rate": 0.0006, + "loss": 4.284032821655273, + "step": 5639 + }, + { + "epoch": 78.3353711790393, + "grad_norm": 0.014272629283368587, + "learning_rate": 0.0006, + "loss": 4.234989166259766, + "step": 5640 + }, + { + "epoch": 78.34934497816595, + "grad_norm": 0.014756826683878899, + "learning_rate": 0.0006, + "loss": 4.143499374389648, + "step": 5641 + }, + { + "epoch": 78.36331877729258, + "grad_norm": 0.016549205407500267, + "learning_rate": 0.0006, + "loss": 4.311034202575684, + "step": 5642 + }, + { + "epoch": 78.37729257641921, + "grad_norm": 0.017735283821821213, + "learning_rate": 0.0006, + "loss": 4.246767520904541, + "step": 5643 + }, + { + "epoch": 78.39126637554585, + "grad_norm": 0.017370613291859627, + "learning_rate": 0.0006, + "loss": 4.2460832595825195, + "step": 5644 + }, + { + "epoch": 78.40524017467249, + "grad_norm": 0.014642846770584583, + "learning_rate": 0.0006, + "loss": 4.303811550140381, + "step": 5645 + }, + { + "epoch": 78.41921397379913, + "grad_norm": 0.013445726595818996, + "learning_rate": 0.0006, + "loss": 4.25921630859375, + "step": 5646 + }, + { + "epoch": 78.43318777292576, + "grad_norm": 0.014254065230488777, + "learning_rate": 0.0006, + "loss": 4.129446506500244, + "step": 5647 + }, + { + "epoch": 78.44716157205241, + "grad_norm": 0.01883019506931305, + "learning_rate": 0.0006, + "loss": 4.191716194152832, + "step": 5648 + }, + { + "epoch": 78.46113537117904, + "grad_norm": 0.0204437468200922, + "learning_rate": 0.0006, + "loss": 4.333085536956787, + "step": 5649 + }, + { + "epoch": 78.47510917030567, + "grad_norm": 0.01652318798005581, + "learning_rate": 0.0006, + "loss": 4.135462760925293, + "step": 5650 + }, + { + "epoch": 78.48908296943232, + "grad_norm": 0.012753551825881004, + "learning_rate": 0.0006, + "loss": 4.195051193237305, + "step": 5651 + }, + { + "epoch": 78.50305676855895, + "grad_norm": 0.01499118935316801, + "learning_rate": 0.0006, + "loss": 4.157092094421387, + "step": 5652 + }, + { + "epoch": 78.5170305676856, + "grad_norm": 0.015383945778012276, + "learning_rate": 0.0006, + "loss": 4.253818035125732, + "step": 5653 + }, + { + "epoch": 78.53100436681223, + "grad_norm": 0.015721332281827927, + "learning_rate": 0.0006, + "loss": 4.228775978088379, + "step": 5654 + }, + { + "epoch": 78.54497816593886, + "grad_norm": 0.014595463871955872, + "learning_rate": 0.0006, + "loss": 4.135197639465332, + "step": 5655 + }, + { + "epoch": 78.5589519650655, + "grad_norm": 0.015059644356369972, + "learning_rate": 0.0006, + "loss": 4.255614280700684, + "step": 5656 + }, + { + "epoch": 78.57292576419214, + "grad_norm": 0.013833344914019108, + "learning_rate": 0.0006, + "loss": 4.301250457763672, + "step": 5657 + }, + { + "epoch": 78.58689956331878, + "grad_norm": 0.013665376231074333, + "learning_rate": 0.0006, + "loss": 4.1734232902526855, + "step": 5658 + }, + { + "epoch": 78.60087336244541, + "grad_norm": 0.01539086177945137, + "learning_rate": 0.0006, + "loss": 4.221459865570068, + "step": 5659 + }, + { + "epoch": 78.61484716157206, + "grad_norm": 0.013313917443156242, + "learning_rate": 0.0006, + "loss": 4.255619049072266, + "step": 5660 + }, + { + "epoch": 78.62882096069869, + "grad_norm": 0.01447741687297821, + "learning_rate": 0.0006, + "loss": 4.298219680786133, + "step": 5661 + }, + { + "epoch": 78.64279475982532, + "grad_norm": 0.012796915136277676, + "learning_rate": 0.0006, + "loss": 4.2542405128479, + "step": 5662 + }, + { + "epoch": 78.65676855895197, + "grad_norm": 0.014058141969144344, + "learning_rate": 0.0006, + "loss": 4.284477233886719, + "step": 5663 + }, + { + "epoch": 78.6707423580786, + "grad_norm": 0.012705069966614246, + "learning_rate": 0.0006, + "loss": 4.2826666831970215, + "step": 5664 + }, + { + "epoch": 78.68471615720524, + "grad_norm": 0.013245890848338604, + "learning_rate": 0.0006, + "loss": 4.371612548828125, + "step": 5665 + }, + { + "epoch": 78.69868995633188, + "grad_norm": 0.014192677102982998, + "learning_rate": 0.0006, + "loss": 4.267344951629639, + "step": 5666 + }, + { + "epoch": 78.71266375545852, + "grad_norm": 0.015346777625381947, + "learning_rate": 0.0006, + "loss": 4.222797393798828, + "step": 5667 + }, + { + "epoch": 78.72663755458515, + "grad_norm": 0.014068621210753918, + "learning_rate": 0.0006, + "loss": 4.172964572906494, + "step": 5668 + }, + { + "epoch": 78.74061135371178, + "grad_norm": 0.014364945702254772, + "learning_rate": 0.0006, + "loss": 4.259319305419922, + "step": 5669 + }, + { + "epoch": 78.75458515283843, + "grad_norm": 0.017337387427687645, + "learning_rate": 0.0006, + "loss": 4.29533576965332, + "step": 5670 + }, + { + "epoch": 78.76855895196506, + "grad_norm": 0.017416715621948242, + "learning_rate": 0.0006, + "loss": 4.340364456176758, + "step": 5671 + }, + { + "epoch": 78.78253275109171, + "grad_norm": 0.015925846993923187, + "learning_rate": 0.0006, + "loss": 4.177454948425293, + "step": 5672 + }, + { + "epoch": 78.79650655021834, + "grad_norm": 0.013130443170666695, + "learning_rate": 0.0006, + "loss": 4.277454853057861, + "step": 5673 + }, + { + "epoch": 78.81048034934499, + "grad_norm": 0.014628687873482704, + "learning_rate": 0.0006, + "loss": 4.273757457733154, + "step": 5674 + }, + { + "epoch": 78.82445414847162, + "grad_norm": 0.01601659320294857, + "learning_rate": 0.0006, + "loss": 4.281774520874023, + "step": 5675 + }, + { + "epoch": 78.83842794759825, + "grad_norm": 0.015443485230207443, + "learning_rate": 0.0006, + "loss": 4.199937343597412, + "step": 5676 + }, + { + "epoch": 78.8524017467249, + "grad_norm": 0.013934246264398098, + "learning_rate": 0.0006, + "loss": 4.331369876861572, + "step": 5677 + }, + { + "epoch": 78.86637554585153, + "grad_norm": 0.01572083681821823, + "learning_rate": 0.0006, + "loss": 4.2828474044799805, + "step": 5678 + }, + { + "epoch": 78.88034934497817, + "grad_norm": 0.013444120064377785, + "learning_rate": 0.0006, + "loss": 4.183859825134277, + "step": 5679 + }, + { + "epoch": 78.8943231441048, + "grad_norm": 0.01591254211962223, + "learning_rate": 0.0006, + "loss": 4.295256614685059, + "step": 5680 + }, + { + "epoch": 78.90829694323143, + "grad_norm": 0.017556684091687202, + "learning_rate": 0.0006, + "loss": 4.132272720336914, + "step": 5681 + }, + { + "epoch": 78.92227074235808, + "grad_norm": 0.018184322863817215, + "learning_rate": 0.0006, + "loss": 4.251728057861328, + "step": 5682 + }, + { + "epoch": 78.93624454148471, + "grad_norm": 0.016207285225391388, + "learning_rate": 0.0006, + "loss": 4.205544471740723, + "step": 5683 + }, + { + "epoch": 78.95021834061136, + "grad_norm": 0.014478376135230064, + "learning_rate": 0.0006, + "loss": 4.182462215423584, + "step": 5684 + }, + { + "epoch": 78.96419213973799, + "grad_norm": 0.014255235902965069, + "learning_rate": 0.0006, + "loss": 4.253366947174072, + "step": 5685 + }, + { + "epoch": 78.97816593886463, + "grad_norm": 0.01524396613240242, + "learning_rate": 0.0006, + "loss": 4.221786022186279, + "step": 5686 + }, + { + "epoch": 78.99213973799127, + "grad_norm": 0.015057693235576153, + "learning_rate": 0.0006, + "loss": 4.208454608917236, + "step": 5687 + }, + { + "epoch": 79.0, + "grad_norm": 0.015478022396564484, + "learning_rate": 0.0006, + "loss": 4.336383819580078, + "step": 5688 + }, + { + "epoch": 79.0, + "eval_loss": 4.641749382019043, + "eval_runtime": 56.8167, + "eval_samples_per_second": 42.98, + "eval_steps_per_second": 1.355, + "step": 5688 + }, + { + "epoch": 79.01397379912663, + "grad_norm": 0.014143316075205803, + "learning_rate": 0.0006, + "loss": 4.260974884033203, + "step": 5689 + }, + { + "epoch": 79.02794759825328, + "grad_norm": 0.013120883144438267, + "learning_rate": 0.0006, + "loss": 4.139886856079102, + "step": 5690 + }, + { + "epoch": 79.04192139737991, + "grad_norm": 0.015019661746919155, + "learning_rate": 0.0006, + "loss": 4.267004013061523, + "step": 5691 + }, + { + "epoch": 79.05589519650655, + "grad_norm": 0.014891284517943859, + "learning_rate": 0.0006, + "loss": 4.205019950866699, + "step": 5692 + }, + { + "epoch": 79.06986899563319, + "grad_norm": 0.015527728945016861, + "learning_rate": 0.0006, + "loss": 4.1513566970825195, + "step": 5693 + }, + { + "epoch": 79.08384279475983, + "grad_norm": 0.017527181655168533, + "learning_rate": 0.0006, + "loss": 4.148828983306885, + "step": 5694 + }, + { + "epoch": 79.09781659388646, + "grad_norm": 0.01632206328213215, + "learning_rate": 0.0006, + "loss": 4.195789813995361, + "step": 5695 + }, + { + "epoch": 79.1117903930131, + "grad_norm": 0.01590503193438053, + "learning_rate": 0.0006, + "loss": 4.199155330657959, + "step": 5696 + }, + { + "epoch": 79.12576419213974, + "grad_norm": 0.016489997506141663, + "learning_rate": 0.0006, + "loss": 4.155154705047607, + "step": 5697 + }, + { + "epoch": 79.13973799126637, + "grad_norm": 0.0161368977278471, + "learning_rate": 0.0006, + "loss": 4.164384841918945, + "step": 5698 + }, + { + "epoch": 79.15371179039302, + "grad_norm": 0.018574809655547142, + "learning_rate": 0.0006, + "loss": 4.1898193359375, + "step": 5699 + }, + { + "epoch": 79.16768558951965, + "grad_norm": 0.018987983465194702, + "learning_rate": 0.0006, + "loss": 4.219719409942627, + "step": 5700 + }, + { + "epoch": 79.18165938864628, + "grad_norm": 0.01668606884777546, + "learning_rate": 0.0006, + "loss": 4.152734756469727, + "step": 5701 + }, + { + "epoch": 79.19563318777293, + "grad_norm": 0.014046299271285534, + "learning_rate": 0.0006, + "loss": 4.152964115142822, + "step": 5702 + }, + { + "epoch": 79.20960698689956, + "grad_norm": 0.01628032699227333, + "learning_rate": 0.0006, + "loss": 4.299479961395264, + "step": 5703 + }, + { + "epoch": 79.2235807860262, + "grad_norm": 0.017998795956373215, + "learning_rate": 0.0006, + "loss": 4.231203079223633, + "step": 5704 + }, + { + "epoch": 79.23755458515284, + "grad_norm": 0.014799647964537144, + "learning_rate": 0.0006, + "loss": 4.3530192375183105, + "step": 5705 + }, + { + "epoch": 79.25152838427948, + "grad_norm": 0.014640678651630878, + "learning_rate": 0.0006, + "loss": 4.287322998046875, + "step": 5706 + }, + { + "epoch": 79.26550218340611, + "grad_norm": 0.0172616895288229, + "learning_rate": 0.0006, + "loss": 4.148009777069092, + "step": 5707 + }, + { + "epoch": 79.27947598253274, + "grad_norm": 0.01811790093779564, + "learning_rate": 0.0006, + "loss": 4.269433975219727, + "step": 5708 + }, + { + "epoch": 79.29344978165939, + "grad_norm": 0.01764000579714775, + "learning_rate": 0.0006, + "loss": 4.292638778686523, + "step": 5709 + }, + { + "epoch": 79.30742358078602, + "grad_norm": 0.018030935898423195, + "learning_rate": 0.0006, + "loss": 4.267889976501465, + "step": 5710 + }, + { + "epoch": 79.32139737991267, + "grad_norm": 0.017860891297459602, + "learning_rate": 0.0006, + "loss": 4.211110591888428, + "step": 5711 + }, + { + "epoch": 79.3353711790393, + "grad_norm": 0.017295073717832565, + "learning_rate": 0.0006, + "loss": 4.224645614624023, + "step": 5712 + }, + { + "epoch": 79.34934497816595, + "grad_norm": 0.015891101211309433, + "learning_rate": 0.0006, + "loss": 4.224797248840332, + "step": 5713 + }, + { + "epoch": 79.36331877729258, + "grad_norm": 0.014680417254567146, + "learning_rate": 0.0006, + "loss": 4.182242393493652, + "step": 5714 + }, + { + "epoch": 79.37729257641921, + "grad_norm": 0.016621140763163567, + "learning_rate": 0.0006, + "loss": 4.186334609985352, + "step": 5715 + }, + { + "epoch": 79.39126637554585, + "grad_norm": 0.016207249835133553, + "learning_rate": 0.0006, + "loss": 4.277074813842773, + "step": 5716 + }, + { + "epoch": 79.40524017467249, + "grad_norm": 0.015375478193163872, + "learning_rate": 0.0006, + "loss": 4.154317855834961, + "step": 5717 + }, + { + "epoch": 79.41921397379913, + "grad_norm": 0.013453228399157524, + "learning_rate": 0.0006, + "loss": 4.265270233154297, + "step": 5718 + }, + { + "epoch": 79.43318777292576, + "grad_norm": 0.014312694780528545, + "learning_rate": 0.0006, + "loss": 4.232220649719238, + "step": 5719 + }, + { + "epoch": 79.44716157205241, + "grad_norm": 0.014759554527699947, + "learning_rate": 0.0006, + "loss": 4.272275924682617, + "step": 5720 + }, + { + "epoch": 79.46113537117904, + "grad_norm": 0.015885505825281143, + "learning_rate": 0.0006, + "loss": 4.2229719161987305, + "step": 5721 + }, + { + "epoch": 79.47510917030567, + "grad_norm": 0.016516581177711487, + "learning_rate": 0.0006, + "loss": 4.200624465942383, + "step": 5722 + }, + { + "epoch": 79.48908296943232, + "grad_norm": 0.018254555761814117, + "learning_rate": 0.0006, + "loss": 4.19648551940918, + "step": 5723 + }, + { + "epoch": 79.50305676855895, + "grad_norm": 0.01467090006917715, + "learning_rate": 0.0006, + "loss": 4.135190486907959, + "step": 5724 + }, + { + "epoch": 79.5170305676856, + "grad_norm": 0.013474797829985619, + "learning_rate": 0.0006, + "loss": 4.311361312866211, + "step": 5725 + }, + { + "epoch": 79.53100436681223, + "grad_norm": 0.01742764748632908, + "learning_rate": 0.0006, + "loss": 4.1845903396606445, + "step": 5726 + }, + { + "epoch": 79.54497816593886, + "grad_norm": 0.018786882981657982, + "learning_rate": 0.0006, + "loss": 4.163942337036133, + "step": 5727 + }, + { + "epoch": 79.5589519650655, + "grad_norm": 0.017831219360232353, + "learning_rate": 0.0006, + "loss": 4.236685276031494, + "step": 5728 + }, + { + "epoch": 79.57292576419214, + "grad_norm": 0.016334986314177513, + "learning_rate": 0.0006, + "loss": 4.255853652954102, + "step": 5729 + }, + { + "epoch": 79.58689956331878, + "grad_norm": 0.0138487434014678, + "learning_rate": 0.0006, + "loss": 4.227874279022217, + "step": 5730 + }, + { + "epoch": 79.60087336244541, + "grad_norm": 0.01531760673969984, + "learning_rate": 0.0006, + "loss": 4.221432685852051, + "step": 5731 + }, + { + "epoch": 79.61484716157206, + "grad_norm": 0.01603361964225769, + "learning_rate": 0.0006, + "loss": 4.278326511383057, + "step": 5732 + }, + { + "epoch": 79.62882096069869, + "grad_norm": 0.013149895705282688, + "learning_rate": 0.0006, + "loss": 4.2474517822265625, + "step": 5733 + }, + { + "epoch": 79.64279475982532, + "grad_norm": 0.013428786769509315, + "learning_rate": 0.0006, + "loss": 4.1701178550720215, + "step": 5734 + }, + { + "epoch": 79.65676855895197, + "grad_norm": 0.014305585995316505, + "learning_rate": 0.0006, + "loss": 4.250778675079346, + "step": 5735 + }, + { + "epoch": 79.6707423580786, + "grad_norm": 0.014544566161930561, + "learning_rate": 0.0006, + "loss": 4.243236541748047, + "step": 5736 + }, + { + "epoch": 79.68471615720524, + "grad_norm": 0.013524414040148258, + "learning_rate": 0.0006, + "loss": 4.255457878112793, + "step": 5737 + }, + { + "epoch": 79.69868995633188, + "grad_norm": 0.014750641770660877, + "learning_rate": 0.0006, + "loss": 4.202816486358643, + "step": 5738 + }, + { + "epoch": 79.71266375545852, + "grad_norm": 0.015090773813426495, + "learning_rate": 0.0006, + "loss": 4.3053178787231445, + "step": 5739 + }, + { + "epoch": 79.72663755458515, + "grad_norm": 0.01606169529259205, + "learning_rate": 0.0006, + "loss": 4.224017143249512, + "step": 5740 + }, + { + "epoch": 79.74061135371178, + "grad_norm": 0.016303520649671555, + "learning_rate": 0.0006, + "loss": 4.190760612487793, + "step": 5741 + }, + { + "epoch": 79.75458515283843, + "grad_norm": 0.01465672068297863, + "learning_rate": 0.0006, + "loss": 4.24506950378418, + "step": 5742 + }, + { + "epoch": 79.76855895196506, + "grad_norm": 0.012616301886737347, + "learning_rate": 0.0006, + "loss": 4.174039840698242, + "step": 5743 + }, + { + "epoch": 79.78253275109171, + "grad_norm": 0.01358284242451191, + "learning_rate": 0.0006, + "loss": 4.236380577087402, + "step": 5744 + }, + { + "epoch": 79.79650655021834, + "grad_norm": 0.01557177770882845, + "learning_rate": 0.0006, + "loss": 4.102962493896484, + "step": 5745 + }, + { + "epoch": 79.81048034934499, + "grad_norm": 0.015584317035973072, + "learning_rate": 0.0006, + "loss": 4.242603302001953, + "step": 5746 + }, + { + "epoch": 79.82445414847162, + "grad_norm": 0.012597334571182728, + "learning_rate": 0.0006, + "loss": 4.202791690826416, + "step": 5747 + }, + { + "epoch": 79.83842794759825, + "grad_norm": 0.013722898438572884, + "learning_rate": 0.0006, + "loss": 4.204328536987305, + "step": 5748 + }, + { + "epoch": 79.8524017467249, + "grad_norm": 0.014669225551187992, + "learning_rate": 0.0006, + "loss": 4.164331912994385, + "step": 5749 + }, + { + "epoch": 79.86637554585153, + "grad_norm": 0.015012740157544613, + "learning_rate": 0.0006, + "loss": 4.260565280914307, + "step": 5750 + }, + { + "epoch": 79.88034934497817, + "grad_norm": 0.016026340425014496, + "learning_rate": 0.0006, + "loss": 4.299723148345947, + "step": 5751 + }, + { + "epoch": 79.8943231441048, + "grad_norm": 0.01626049168407917, + "learning_rate": 0.0006, + "loss": 4.207252502441406, + "step": 5752 + }, + { + "epoch": 79.90829694323143, + "grad_norm": 0.015415354631841183, + "learning_rate": 0.0006, + "loss": 4.151934623718262, + "step": 5753 + }, + { + "epoch": 79.92227074235808, + "grad_norm": 0.014158242382109165, + "learning_rate": 0.0006, + "loss": 4.3461995124816895, + "step": 5754 + }, + { + "epoch": 79.93624454148471, + "grad_norm": 0.01585628278553486, + "learning_rate": 0.0006, + "loss": 4.334761619567871, + "step": 5755 + }, + { + "epoch": 79.95021834061136, + "grad_norm": 0.013710759580135345, + "learning_rate": 0.0006, + "loss": 4.180401802062988, + "step": 5756 + }, + { + "epoch": 79.96419213973799, + "grad_norm": 0.014159204438328743, + "learning_rate": 0.0006, + "loss": 4.322331428527832, + "step": 5757 + }, + { + "epoch": 79.97816593886463, + "grad_norm": 0.015153195708990097, + "learning_rate": 0.0006, + "loss": 4.287478923797607, + "step": 5758 + }, + { + "epoch": 79.99213973799127, + "grad_norm": 0.014100227504968643, + "learning_rate": 0.0006, + "loss": 4.231807708740234, + "step": 5759 + }, + { + "epoch": 80.0, + "grad_norm": 0.015248682349920273, + "learning_rate": 0.0006, + "loss": 4.261106014251709, + "step": 5760 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.455690689065779e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-5760/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-5760/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-5760/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6048/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6048/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ad65be1b697f018928914a93b854e0496abad24 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc2cdfa23e858d40673eea21eddfd41cda0e7275b967ad1d7cc7cfc978436642 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6048/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3dcba1342bf1b37a3167694c7a637b82ea650136 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02943afd569dead523bc35a26b6f3fd542955ac7b567d4ef39df4b1aa8c2d59a +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..312cfda1b3b5a4c030872e90729af7fda6414a64 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1488a6864d0d399732d11ecf9366384d524ff67d5ccdb266991648ccbfd7b9bc +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..015687cd7430dbe642bb8b7450a399b412a88c06 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:155a2a845fd8d4bda671f5716878c0564ba65e425b4af3efa44ce4c4b04e3988 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..329ea246992b66631e9ecb4e675f05d70753f65b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79f726cea75304e98d91d809622dae58023fec445e7eeed8e7c7165af5871843 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..63952880ff4ff8278c9056ac3378e4fd623b5e74 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a2073d8191537a847cb1080353e4ac586e2f8357bba6e7741b41fb002ce8e07 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6048/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..762789ce7e31cc09dc831dc70530516d1f1e01b2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:100b47a764841837311a60bed293b8e5dd3c3501d4d1da6c3b20120bb6630ddd +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6048/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5369c2ee5a37252b276cd1b917005258493c9dcf --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/trainer_state.json @@ -0,0 +1,43034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 84.0, + "eval_steps": 500, + "global_step": 6048, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + }, + { + "epoch": 76.0, + "eval_loss": 4.689640998840332, + "eval_runtime": 56.9095, + "eval_samples_per_second": 42.91, + "eval_steps_per_second": 1.353, + "step": 5472 + }, + { + "epoch": 76.01397379912663, + "grad_norm": 0.015351055189967155, + "learning_rate": 0.0006, + "loss": 4.348268985748291, + "step": 5473 + }, + { + "epoch": 76.02794759825328, + "grad_norm": 0.017164217308163643, + "learning_rate": 0.0006, + "loss": 4.244204998016357, + "step": 5474 + }, + { + "epoch": 76.04192139737991, + "grad_norm": 0.021624932065606117, + "learning_rate": 0.0006, + "loss": 4.311168670654297, + "step": 5475 + }, + { + "epoch": 76.05589519650655, + "grad_norm": 0.028712468221783638, + "learning_rate": 0.0006, + "loss": 4.322287082672119, + "step": 5476 + }, + { + "epoch": 76.06986899563319, + "grad_norm": 0.027915487065911293, + "learning_rate": 0.0006, + "loss": 4.381207466125488, + "step": 5477 + }, + { + "epoch": 76.08384279475983, + "grad_norm": 0.017955303192138672, + "learning_rate": 0.0006, + "loss": 4.2635016441345215, + "step": 5478 + }, + { + "epoch": 76.09781659388646, + "grad_norm": 0.019113514572381973, + "learning_rate": 0.0006, + "loss": 4.399471282958984, + "step": 5479 + }, + { + "epoch": 76.1117903930131, + "grad_norm": 0.017502542585134506, + "learning_rate": 0.0006, + "loss": 4.321204662322998, + "step": 5480 + }, + { + "epoch": 76.12576419213974, + "grad_norm": 0.01852525770664215, + "learning_rate": 0.0006, + "loss": 4.217884063720703, + "step": 5481 + }, + { + "epoch": 76.13973799126637, + "grad_norm": 0.019836828112602234, + "learning_rate": 0.0006, + "loss": 4.278532028198242, + "step": 5482 + }, + { + "epoch": 76.15371179039302, + "grad_norm": 0.02101975679397583, + "learning_rate": 0.0006, + "loss": 4.246720314025879, + "step": 5483 + }, + { + "epoch": 76.16768558951965, + "grad_norm": 0.021202990785241127, + "learning_rate": 0.0006, + "loss": 4.517517566680908, + "step": 5484 + }, + { + "epoch": 76.18165938864628, + "grad_norm": 0.021428769454360008, + "learning_rate": 0.0006, + "loss": 4.3373517990112305, + "step": 5485 + }, + { + "epoch": 76.19563318777293, + "grad_norm": 0.018263190984725952, + "learning_rate": 0.0006, + "loss": 4.402856826782227, + "step": 5486 + }, + { + "epoch": 76.20960698689956, + "grad_norm": 0.019025536254048347, + "learning_rate": 0.0006, + "loss": 4.314227104187012, + "step": 5487 + }, + { + "epoch": 76.2235807860262, + "grad_norm": 0.020918630063533783, + "learning_rate": 0.0006, + "loss": 4.454369068145752, + "step": 5488 + }, + { + "epoch": 76.23755458515284, + "grad_norm": 0.019307617098093033, + "learning_rate": 0.0006, + "loss": 4.247509002685547, + "step": 5489 + }, + { + "epoch": 76.25152838427948, + "grad_norm": 0.014956331811845303, + "learning_rate": 0.0006, + "loss": 4.365175247192383, + "step": 5490 + }, + { + "epoch": 76.26550218340611, + "grad_norm": 0.017213059589266777, + "learning_rate": 0.0006, + "loss": 4.384005069732666, + "step": 5491 + }, + { + "epoch": 76.27947598253274, + "grad_norm": 0.017575597390532494, + "learning_rate": 0.0006, + "loss": 4.41934871673584, + "step": 5492 + }, + { + "epoch": 76.29344978165939, + "grad_norm": 0.015879256650805473, + "learning_rate": 0.0006, + "loss": 4.26675271987915, + "step": 5493 + }, + { + "epoch": 76.30742358078602, + "grad_norm": 0.01564522460103035, + "learning_rate": 0.0006, + "loss": 4.277339935302734, + "step": 5494 + }, + { + "epoch": 76.32139737991267, + "grad_norm": 0.016414586454629898, + "learning_rate": 0.0006, + "loss": 4.4138593673706055, + "step": 5495 + }, + { + "epoch": 76.3353711790393, + "grad_norm": 0.014732337556779385, + "learning_rate": 0.0006, + "loss": 4.281006336212158, + "step": 5496 + }, + { + "epoch": 76.34934497816595, + "grad_norm": 0.015461455099284649, + "learning_rate": 0.0006, + "loss": 4.263422966003418, + "step": 5497 + }, + { + "epoch": 76.36331877729258, + "grad_norm": 0.014529814012348652, + "learning_rate": 0.0006, + "loss": 4.36165189743042, + "step": 5498 + }, + { + "epoch": 76.37729257641921, + "grad_norm": 0.01469483319669962, + "learning_rate": 0.0006, + "loss": 4.309642791748047, + "step": 5499 + }, + { + "epoch": 76.39126637554585, + "grad_norm": 0.016490574926137924, + "learning_rate": 0.0006, + "loss": 4.288150787353516, + "step": 5500 + }, + { + "epoch": 76.40524017467249, + "grad_norm": 0.014593565836548805, + "learning_rate": 0.0006, + "loss": 4.271614074707031, + "step": 5501 + }, + { + "epoch": 76.41921397379913, + "grad_norm": 0.012796535156667233, + "learning_rate": 0.0006, + "loss": 4.420042037963867, + "step": 5502 + }, + { + "epoch": 76.43318777292576, + "grad_norm": 0.016086025163531303, + "learning_rate": 0.0006, + "loss": 4.289943695068359, + "step": 5503 + }, + { + "epoch": 76.44716157205241, + "grad_norm": 0.015439470298588276, + "learning_rate": 0.0006, + "loss": 4.3023271560668945, + "step": 5504 + }, + { + "epoch": 76.46113537117904, + "grad_norm": 0.015066367574036121, + "learning_rate": 0.0006, + "loss": 4.197081565856934, + "step": 5505 + }, + { + "epoch": 76.47510917030567, + "grad_norm": 0.016653263941407204, + "learning_rate": 0.0006, + "loss": 4.333785057067871, + "step": 5506 + }, + { + "epoch": 76.48908296943232, + "grad_norm": 0.014791973866522312, + "learning_rate": 0.0006, + "loss": 4.323811054229736, + "step": 5507 + }, + { + "epoch": 76.50305676855895, + "grad_norm": 0.014694827608764172, + "learning_rate": 0.0006, + "loss": 4.195852279663086, + "step": 5508 + }, + { + "epoch": 76.5170305676856, + "grad_norm": 0.015120396390557289, + "learning_rate": 0.0006, + "loss": 4.42379903793335, + "step": 5509 + }, + { + "epoch": 76.53100436681223, + "grad_norm": 0.014584025368094444, + "learning_rate": 0.0006, + "loss": 4.2702531814575195, + "step": 5510 + }, + { + "epoch": 76.54497816593886, + "grad_norm": 0.013981613330543041, + "learning_rate": 0.0006, + "loss": 4.383029937744141, + "step": 5511 + }, + { + "epoch": 76.5589519650655, + "grad_norm": 0.012930642813444138, + "learning_rate": 0.0006, + "loss": 4.296136856079102, + "step": 5512 + }, + { + "epoch": 76.57292576419214, + "grad_norm": 0.013060989789664745, + "learning_rate": 0.0006, + "loss": 4.346380233764648, + "step": 5513 + }, + { + "epoch": 76.58689956331878, + "grad_norm": 0.013209798373281956, + "learning_rate": 0.0006, + "loss": 4.316067218780518, + "step": 5514 + }, + { + "epoch": 76.60087336244541, + "grad_norm": 0.012423216365277767, + "learning_rate": 0.0006, + "loss": 4.370274543762207, + "step": 5515 + }, + { + "epoch": 76.61484716157206, + "grad_norm": 0.013205411843955517, + "learning_rate": 0.0006, + "loss": 4.429969310760498, + "step": 5516 + }, + { + "epoch": 76.62882096069869, + "grad_norm": 0.013049827888607979, + "learning_rate": 0.0006, + "loss": 4.378549098968506, + "step": 5517 + }, + { + "epoch": 76.64279475982532, + "grad_norm": 0.012075880542397499, + "learning_rate": 0.0006, + "loss": 4.348293304443359, + "step": 5518 + }, + { + "epoch": 76.65676855895197, + "grad_norm": 0.014932847581803799, + "learning_rate": 0.0006, + "loss": 4.282538890838623, + "step": 5519 + }, + { + "epoch": 76.6707423580786, + "grad_norm": 0.015615998767316341, + "learning_rate": 0.0006, + "loss": 4.274641036987305, + "step": 5520 + }, + { + "epoch": 76.68471615720524, + "grad_norm": 0.01424513477832079, + "learning_rate": 0.0006, + "loss": 4.243166923522949, + "step": 5521 + }, + { + "epoch": 76.69868995633188, + "grad_norm": 0.012775546871125698, + "learning_rate": 0.0006, + "loss": 4.335611820220947, + "step": 5522 + }, + { + "epoch": 76.71266375545852, + "grad_norm": 0.014461303129792213, + "learning_rate": 0.0006, + "loss": 4.246976375579834, + "step": 5523 + }, + { + "epoch": 76.72663755458515, + "grad_norm": 0.0145436255261302, + "learning_rate": 0.0006, + "loss": 4.275895118713379, + "step": 5524 + }, + { + "epoch": 76.74061135371178, + "grad_norm": 0.014055085368454456, + "learning_rate": 0.0006, + "loss": 4.331108093261719, + "step": 5525 + }, + { + "epoch": 76.75458515283843, + "grad_norm": 0.011933894827961922, + "learning_rate": 0.0006, + "loss": 4.293721675872803, + "step": 5526 + }, + { + "epoch": 76.76855895196506, + "grad_norm": 0.01277392078191042, + "learning_rate": 0.0006, + "loss": 4.312374114990234, + "step": 5527 + }, + { + "epoch": 76.78253275109171, + "grad_norm": 0.014227936044335365, + "learning_rate": 0.0006, + "loss": 4.238729000091553, + "step": 5528 + }, + { + "epoch": 76.79650655021834, + "grad_norm": 0.01488024927675724, + "learning_rate": 0.0006, + "loss": 4.24587345123291, + "step": 5529 + }, + { + "epoch": 76.81048034934499, + "grad_norm": 0.015341250225901604, + "learning_rate": 0.0006, + "loss": 4.257378578186035, + "step": 5530 + }, + { + "epoch": 76.82445414847162, + "grad_norm": 0.015179437585175037, + "learning_rate": 0.0006, + "loss": 4.2812910079956055, + "step": 5531 + }, + { + "epoch": 76.83842794759825, + "grad_norm": 0.012814724817872047, + "learning_rate": 0.0006, + "loss": 4.2309465408325195, + "step": 5532 + }, + { + "epoch": 76.8524017467249, + "grad_norm": 0.013810716569423676, + "learning_rate": 0.0006, + "loss": 4.325292110443115, + "step": 5533 + }, + { + "epoch": 76.86637554585153, + "grad_norm": 0.017468232661485672, + "learning_rate": 0.0006, + "loss": 4.243788719177246, + "step": 5534 + }, + { + "epoch": 76.88034934497817, + "grad_norm": 0.01739857904613018, + "learning_rate": 0.0006, + "loss": 4.345390796661377, + "step": 5535 + }, + { + "epoch": 76.8943231441048, + "grad_norm": 0.015329061076045036, + "learning_rate": 0.0006, + "loss": 4.415142059326172, + "step": 5536 + }, + { + "epoch": 76.90829694323143, + "grad_norm": 0.01598426327109337, + "learning_rate": 0.0006, + "loss": 4.299558639526367, + "step": 5537 + }, + { + "epoch": 76.92227074235808, + "grad_norm": 0.01621454581618309, + "learning_rate": 0.0006, + "loss": 4.284354209899902, + "step": 5538 + }, + { + "epoch": 76.93624454148471, + "grad_norm": 0.015956968069076538, + "learning_rate": 0.0006, + "loss": 4.293837547302246, + "step": 5539 + }, + { + "epoch": 76.95021834061136, + "grad_norm": 0.01776493526995182, + "learning_rate": 0.0006, + "loss": 4.229729175567627, + "step": 5540 + }, + { + "epoch": 76.96419213973799, + "grad_norm": 0.015317477285861969, + "learning_rate": 0.0006, + "loss": 4.3328447341918945, + "step": 5541 + }, + { + "epoch": 76.97816593886463, + "grad_norm": 0.015140251256525517, + "learning_rate": 0.0006, + "loss": 4.310840606689453, + "step": 5542 + }, + { + "epoch": 76.99213973799127, + "grad_norm": 0.015326812863349915, + "learning_rate": 0.0006, + "loss": 4.230832099914551, + "step": 5543 + }, + { + "epoch": 77.0, + "grad_norm": 0.01607763208448887, + "learning_rate": 0.0006, + "loss": 4.2236328125, + "step": 5544 + }, + { + "epoch": 77.0, + "eval_loss": 4.662221431732178, + "eval_runtime": 56.5194, + "eval_samples_per_second": 43.206, + "eval_steps_per_second": 1.362, + "step": 5544 + }, + { + "epoch": 77.01397379912663, + "grad_norm": 0.014839510433375835, + "learning_rate": 0.0006, + "loss": 4.254111289978027, + "step": 5545 + }, + { + "epoch": 77.02794759825328, + "grad_norm": 0.01519741676747799, + "learning_rate": 0.0006, + "loss": 4.1607184410095215, + "step": 5546 + }, + { + "epoch": 77.04192139737991, + "grad_norm": 0.015472603030502796, + "learning_rate": 0.0006, + "loss": 4.323320388793945, + "step": 5547 + }, + { + "epoch": 77.05589519650655, + "grad_norm": 0.017937948927283287, + "learning_rate": 0.0006, + "loss": 4.308490753173828, + "step": 5548 + }, + { + "epoch": 77.06986899563319, + "grad_norm": 0.020890841260552406, + "learning_rate": 0.0006, + "loss": 4.259101867675781, + "step": 5549 + }, + { + "epoch": 77.08384279475983, + "grad_norm": 0.01892857998609543, + "learning_rate": 0.0006, + "loss": 4.386970520019531, + "step": 5550 + }, + { + "epoch": 77.09781659388646, + "grad_norm": 0.016583630815148354, + "learning_rate": 0.0006, + "loss": 4.233604907989502, + "step": 5551 + }, + { + "epoch": 77.1117903930131, + "grad_norm": 0.016063468530774117, + "learning_rate": 0.0006, + "loss": 4.339544296264648, + "step": 5552 + }, + { + "epoch": 77.12576419213974, + "grad_norm": 0.018326779827475548, + "learning_rate": 0.0006, + "loss": 4.367204666137695, + "step": 5553 + }, + { + "epoch": 77.13973799126637, + "grad_norm": 0.021590933203697205, + "learning_rate": 0.0006, + "loss": 4.352685928344727, + "step": 5554 + }, + { + "epoch": 77.15371179039302, + "grad_norm": 0.019643642008304596, + "learning_rate": 0.0006, + "loss": 4.258535385131836, + "step": 5555 + }, + { + "epoch": 77.16768558951965, + "grad_norm": 0.018476149067282677, + "learning_rate": 0.0006, + "loss": 4.232568740844727, + "step": 5556 + }, + { + "epoch": 77.18165938864628, + "grad_norm": 0.01584405079483986, + "learning_rate": 0.0006, + "loss": 4.258371353149414, + "step": 5557 + }, + { + "epoch": 77.19563318777293, + "grad_norm": 0.01575201191008091, + "learning_rate": 0.0006, + "loss": 4.183218002319336, + "step": 5558 + }, + { + "epoch": 77.20960698689956, + "grad_norm": 0.016871824860572815, + "learning_rate": 0.0006, + "loss": 4.194563865661621, + "step": 5559 + }, + { + "epoch": 77.2235807860262, + "grad_norm": 0.02059333398938179, + "learning_rate": 0.0006, + "loss": 4.195281982421875, + "step": 5560 + }, + { + "epoch": 77.23755458515284, + "grad_norm": 0.022640453651547432, + "learning_rate": 0.0006, + "loss": 4.279615879058838, + "step": 5561 + }, + { + "epoch": 77.25152838427948, + "grad_norm": 0.020335959270596504, + "learning_rate": 0.0006, + "loss": 4.25040864944458, + "step": 5562 + }, + { + "epoch": 77.26550218340611, + "grad_norm": 0.020010793581604958, + "learning_rate": 0.0006, + "loss": 4.291116714477539, + "step": 5563 + }, + { + "epoch": 77.27947598253274, + "grad_norm": 0.02109045907855034, + "learning_rate": 0.0006, + "loss": 4.364088535308838, + "step": 5564 + }, + { + "epoch": 77.29344978165939, + "grad_norm": 0.019519446417689323, + "learning_rate": 0.0006, + "loss": 4.329166412353516, + "step": 5565 + }, + { + "epoch": 77.30742358078602, + "grad_norm": 0.01689804159104824, + "learning_rate": 0.0006, + "loss": 4.244176864624023, + "step": 5566 + }, + { + "epoch": 77.32139737991267, + "grad_norm": 0.01893681474030018, + "learning_rate": 0.0006, + "loss": 4.239068984985352, + "step": 5567 + }, + { + "epoch": 77.3353711790393, + "grad_norm": 0.015998193994164467, + "learning_rate": 0.0006, + "loss": 4.173398017883301, + "step": 5568 + }, + { + "epoch": 77.34934497816595, + "grad_norm": 0.013892280869185925, + "learning_rate": 0.0006, + "loss": 4.287654876708984, + "step": 5569 + }, + { + "epoch": 77.36331877729258, + "grad_norm": 0.017060643061995506, + "learning_rate": 0.0006, + "loss": 4.398447036743164, + "step": 5570 + }, + { + "epoch": 77.37729257641921, + "grad_norm": 0.015168849378824234, + "learning_rate": 0.0006, + "loss": 4.15665864944458, + "step": 5571 + }, + { + "epoch": 77.39126637554585, + "grad_norm": 0.014886523596942425, + "learning_rate": 0.0006, + "loss": 4.196673393249512, + "step": 5572 + }, + { + "epoch": 77.40524017467249, + "grad_norm": 0.016941796988248825, + "learning_rate": 0.0006, + "loss": 4.314562797546387, + "step": 5573 + }, + { + "epoch": 77.41921397379913, + "grad_norm": 0.015353812836110592, + "learning_rate": 0.0006, + "loss": 4.409579277038574, + "step": 5574 + }, + { + "epoch": 77.43318777292576, + "grad_norm": 0.015192852355539799, + "learning_rate": 0.0006, + "loss": 4.392045974731445, + "step": 5575 + }, + { + "epoch": 77.44716157205241, + "grad_norm": 0.014232569374144077, + "learning_rate": 0.0006, + "loss": 4.21639347076416, + "step": 5576 + }, + { + "epoch": 77.46113537117904, + "grad_norm": 0.014453226700425148, + "learning_rate": 0.0006, + "loss": 4.2802276611328125, + "step": 5577 + }, + { + "epoch": 77.47510917030567, + "grad_norm": 0.012881236150860786, + "learning_rate": 0.0006, + "loss": 4.375980854034424, + "step": 5578 + }, + { + "epoch": 77.48908296943232, + "grad_norm": 0.01583976298570633, + "learning_rate": 0.0006, + "loss": 4.264520645141602, + "step": 5579 + }, + { + "epoch": 77.50305676855895, + "grad_norm": 0.014171584509313107, + "learning_rate": 0.0006, + "loss": 4.225279808044434, + "step": 5580 + }, + { + "epoch": 77.5170305676856, + "grad_norm": 0.014715159311890602, + "learning_rate": 0.0006, + "loss": 4.369834899902344, + "step": 5581 + }, + { + "epoch": 77.53100436681223, + "grad_norm": 0.016554079949855804, + "learning_rate": 0.0006, + "loss": 4.2976179122924805, + "step": 5582 + }, + { + "epoch": 77.54497816593886, + "grad_norm": 0.015095747075974941, + "learning_rate": 0.0006, + "loss": 4.294407844543457, + "step": 5583 + }, + { + "epoch": 77.5589519650655, + "grad_norm": 0.012726361863315105, + "learning_rate": 0.0006, + "loss": 4.221401214599609, + "step": 5584 + }, + { + "epoch": 77.57292576419214, + "grad_norm": 0.014999309554696083, + "learning_rate": 0.0006, + "loss": 4.243929386138916, + "step": 5585 + }, + { + "epoch": 77.58689956331878, + "grad_norm": 0.016526037827134132, + "learning_rate": 0.0006, + "loss": 4.374139785766602, + "step": 5586 + }, + { + "epoch": 77.60087336244541, + "grad_norm": 0.01648339442908764, + "learning_rate": 0.0006, + "loss": 4.392425060272217, + "step": 5587 + }, + { + "epoch": 77.61484716157206, + "grad_norm": 0.015269213356077671, + "learning_rate": 0.0006, + "loss": 4.33986759185791, + "step": 5588 + }, + { + "epoch": 77.62882096069869, + "grad_norm": 0.01690507121384144, + "learning_rate": 0.0006, + "loss": 4.307476043701172, + "step": 5589 + }, + { + "epoch": 77.64279475982532, + "grad_norm": 0.016438685357570648, + "learning_rate": 0.0006, + "loss": 4.2406206130981445, + "step": 5590 + }, + { + "epoch": 77.65676855895197, + "grad_norm": 0.0157600287348032, + "learning_rate": 0.0006, + "loss": 4.314070224761963, + "step": 5591 + }, + { + "epoch": 77.6707423580786, + "grad_norm": 0.015127947553992271, + "learning_rate": 0.0006, + "loss": 4.22475528717041, + "step": 5592 + }, + { + "epoch": 77.68471615720524, + "grad_norm": 0.016632677987217903, + "learning_rate": 0.0006, + "loss": 4.26954984664917, + "step": 5593 + }, + { + "epoch": 77.69868995633188, + "grad_norm": 0.013150460086762905, + "learning_rate": 0.0006, + "loss": 4.369131088256836, + "step": 5594 + }, + { + "epoch": 77.71266375545852, + "grad_norm": 0.014513005502521992, + "learning_rate": 0.0006, + "loss": 4.3248796463012695, + "step": 5595 + }, + { + "epoch": 77.72663755458515, + "grad_norm": 0.014090991578996181, + "learning_rate": 0.0006, + "loss": 4.268381118774414, + "step": 5596 + }, + { + "epoch": 77.74061135371178, + "grad_norm": 0.01633097417652607, + "learning_rate": 0.0006, + "loss": 4.238439083099365, + "step": 5597 + }, + { + "epoch": 77.75458515283843, + "grad_norm": 0.01710933819413185, + "learning_rate": 0.0006, + "loss": 4.399697780609131, + "step": 5598 + }, + { + "epoch": 77.76855895196506, + "grad_norm": 0.019251767545938492, + "learning_rate": 0.0006, + "loss": 4.290450096130371, + "step": 5599 + }, + { + "epoch": 77.78253275109171, + "grad_norm": 0.01849171705543995, + "learning_rate": 0.0006, + "loss": 4.267734527587891, + "step": 5600 + }, + { + "epoch": 77.79650655021834, + "grad_norm": 0.014173115603625774, + "learning_rate": 0.0006, + "loss": 4.288799285888672, + "step": 5601 + }, + { + "epoch": 77.81048034934499, + "grad_norm": 0.014572718180716038, + "learning_rate": 0.0006, + "loss": 4.3790435791015625, + "step": 5602 + }, + { + "epoch": 77.82445414847162, + "grad_norm": 0.016375336796045303, + "learning_rate": 0.0006, + "loss": 4.231725692749023, + "step": 5603 + }, + { + "epoch": 77.83842794759825, + "grad_norm": 0.017326701432466507, + "learning_rate": 0.0006, + "loss": 4.211495399475098, + "step": 5604 + }, + { + "epoch": 77.8524017467249, + "grad_norm": 0.016399390995502472, + "learning_rate": 0.0006, + "loss": 4.386699199676514, + "step": 5605 + }, + { + "epoch": 77.86637554585153, + "grad_norm": 0.018269989639520645, + "learning_rate": 0.0006, + "loss": 4.254856109619141, + "step": 5606 + }, + { + "epoch": 77.88034934497817, + "grad_norm": 0.018443435430526733, + "learning_rate": 0.0006, + "loss": 4.391961097717285, + "step": 5607 + }, + { + "epoch": 77.8943231441048, + "grad_norm": 0.01784636080265045, + "learning_rate": 0.0006, + "loss": 4.273881912231445, + "step": 5608 + }, + { + "epoch": 77.90829694323143, + "grad_norm": 0.01782994158565998, + "learning_rate": 0.0006, + "loss": 4.26833963394165, + "step": 5609 + }, + { + "epoch": 77.92227074235808, + "grad_norm": 0.016260741278529167, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 5610 + }, + { + "epoch": 77.93624454148471, + "grad_norm": 0.015482903458178043, + "learning_rate": 0.0006, + "loss": 4.187826633453369, + "step": 5611 + }, + { + "epoch": 77.95021834061136, + "grad_norm": 0.015193257480859756, + "learning_rate": 0.0006, + "loss": 4.286060333251953, + "step": 5612 + }, + { + "epoch": 77.96419213973799, + "grad_norm": 0.015110115520656109, + "learning_rate": 0.0006, + "loss": 4.2828192710876465, + "step": 5613 + }, + { + "epoch": 77.97816593886463, + "grad_norm": 0.015973486006259918, + "learning_rate": 0.0006, + "loss": 4.317939758300781, + "step": 5614 + }, + { + "epoch": 77.99213973799127, + "grad_norm": 0.017363594844937325, + "learning_rate": 0.0006, + "loss": 4.287721633911133, + "step": 5615 + }, + { + "epoch": 78.0, + "grad_norm": 0.015277642756700516, + "learning_rate": 0.0006, + "loss": 4.356618881225586, + "step": 5616 + }, + { + "epoch": 78.0, + "eval_loss": 4.642678737640381, + "eval_runtime": 56.7335, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 5616 + }, + { + "epoch": 78.01397379912663, + "grad_norm": 0.016794051975011826, + "learning_rate": 0.0006, + "loss": 4.279930591583252, + "step": 5617 + }, + { + "epoch": 78.02794759825328, + "grad_norm": 0.01632673107087612, + "learning_rate": 0.0006, + "loss": 4.235943794250488, + "step": 5618 + }, + { + "epoch": 78.04192139737991, + "grad_norm": 0.01437876746058464, + "learning_rate": 0.0006, + "loss": 4.2542243003845215, + "step": 5619 + }, + { + "epoch": 78.05589519650655, + "grad_norm": 0.014993119053542614, + "learning_rate": 0.0006, + "loss": 4.262046813964844, + "step": 5620 + }, + { + "epoch": 78.06986899563319, + "grad_norm": 0.015721119940280914, + "learning_rate": 0.0006, + "loss": 4.173436164855957, + "step": 5621 + }, + { + "epoch": 78.08384279475983, + "grad_norm": 0.01334085687994957, + "learning_rate": 0.0006, + "loss": 4.322175979614258, + "step": 5622 + }, + { + "epoch": 78.09781659388646, + "grad_norm": 0.015561388805508614, + "learning_rate": 0.0006, + "loss": 4.242125988006592, + "step": 5623 + }, + { + "epoch": 78.1117903930131, + "grad_norm": 0.014418602921068668, + "learning_rate": 0.0006, + "loss": 4.266186714172363, + "step": 5624 + }, + { + "epoch": 78.12576419213974, + "grad_norm": 0.013739909045398235, + "learning_rate": 0.0006, + "loss": 4.231633186340332, + "step": 5625 + }, + { + "epoch": 78.13973799126637, + "grad_norm": 0.01402067206799984, + "learning_rate": 0.0006, + "loss": 4.188406944274902, + "step": 5626 + }, + { + "epoch": 78.15371179039302, + "grad_norm": 0.013639481738209724, + "learning_rate": 0.0006, + "loss": 4.1586594581604, + "step": 5627 + }, + { + "epoch": 78.16768558951965, + "grad_norm": 0.012858159840106964, + "learning_rate": 0.0006, + "loss": 4.360448360443115, + "step": 5628 + }, + { + "epoch": 78.18165938864628, + "grad_norm": 0.013089739717543125, + "learning_rate": 0.0006, + "loss": 4.344371795654297, + "step": 5629 + }, + { + "epoch": 78.19563318777293, + "grad_norm": 0.01382314320653677, + "learning_rate": 0.0006, + "loss": 4.20378303527832, + "step": 5630 + }, + { + "epoch": 78.20960698689956, + "grad_norm": 0.014206619001924992, + "learning_rate": 0.0006, + "loss": 4.239891529083252, + "step": 5631 + }, + { + "epoch": 78.2235807860262, + "grad_norm": 0.01472168043255806, + "learning_rate": 0.0006, + "loss": 4.282382488250732, + "step": 5632 + }, + { + "epoch": 78.23755458515284, + "grad_norm": 0.014237025752663612, + "learning_rate": 0.0006, + "loss": 4.241545677185059, + "step": 5633 + }, + { + "epoch": 78.25152838427948, + "grad_norm": 0.013231472112238407, + "learning_rate": 0.0006, + "loss": 4.184896469116211, + "step": 5634 + }, + { + "epoch": 78.26550218340611, + "grad_norm": 0.013538471423089504, + "learning_rate": 0.0006, + "loss": 4.325181007385254, + "step": 5635 + }, + { + "epoch": 78.27947598253274, + "grad_norm": 0.013541961088776588, + "learning_rate": 0.0006, + "loss": 4.368722915649414, + "step": 5636 + }, + { + "epoch": 78.29344978165939, + "grad_norm": 0.013446723110973835, + "learning_rate": 0.0006, + "loss": 4.301422595977783, + "step": 5637 + }, + { + "epoch": 78.30742358078602, + "grad_norm": 0.014753809198737144, + "learning_rate": 0.0006, + "loss": 4.259127616882324, + "step": 5638 + }, + { + "epoch": 78.32139737991267, + "grad_norm": 0.013256008736789227, + "learning_rate": 0.0006, + "loss": 4.284032821655273, + "step": 5639 + }, + { + "epoch": 78.3353711790393, + "grad_norm": 0.014272629283368587, + "learning_rate": 0.0006, + "loss": 4.234989166259766, + "step": 5640 + }, + { + "epoch": 78.34934497816595, + "grad_norm": 0.014756826683878899, + "learning_rate": 0.0006, + "loss": 4.143499374389648, + "step": 5641 + }, + { + "epoch": 78.36331877729258, + "grad_norm": 0.016549205407500267, + "learning_rate": 0.0006, + "loss": 4.311034202575684, + "step": 5642 + }, + { + "epoch": 78.37729257641921, + "grad_norm": 0.017735283821821213, + "learning_rate": 0.0006, + "loss": 4.246767520904541, + "step": 5643 + }, + { + "epoch": 78.39126637554585, + "grad_norm": 0.017370613291859627, + "learning_rate": 0.0006, + "loss": 4.2460832595825195, + "step": 5644 + }, + { + "epoch": 78.40524017467249, + "grad_norm": 0.014642846770584583, + "learning_rate": 0.0006, + "loss": 4.303811550140381, + "step": 5645 + }, + { + "epoch": 78.41921397379913, + "grad_norm": 0.013445726595818996, + "learning_rate": 0.0006, + "loss": 4.25921630859375, + "step": 5646 + }, + { + "epoch": 78.43318777292576, + "grad_norm": 0.014254065230488777, + "learning_rate": 0.0006, + "loss": 4.129446506500244, + "step": 5647 + }, + { + "epoch": 78.44716157205241, + "grad_norm": 0.01883019506931305, + "learning_rate": 0.0006, + "loss": 4.191716194152832, + "step": 5648 + }, + { + "epoch": 78.46113537117904, + "grad_norm": 0.0204437468200922, + "learning_rate": 0.0006, + "loss": 4.333085536956787, + "step": 5649 + }, + { + "epoch": 78.47510917030567, + "grad_norm": 0.01652318798005581, + "learning_rate": 0.0006, + "loss": 4.135462760925293, + "step": 5650 + }, + { + "epoch": 78.48908296943232, + "grad_norm": 0.012753551825881004, + "learning_rate": 0.0006, + "loss": 4.195051193237305, + "step": 5651 + }, + { + "epoch": 78.50305676855895, + "grad_norm": 0.01499118935316801, + "learning_rate": 0.0006, + "loss": 4.157092094421387, + "step": 5652 + }, + { + "epoch": 78.5170305676856, + "grad_norm": 0.015383945778012276, + "learning_rate": 0.0006, + "loss": 4.253818035125732, + "step": 5653 + }, + { + "epoch": 78.53100436681223, + "grad_norm": 0.015721332281827927, + "learning_rate": 0.0006, + "loss": 4.228775978088379, + "step": 5654 + }, + { + "epoch": 78.54497816593886, + "grad_norm": 0.014595463871955872, + "learning_rate": 0.0006, + "loss": 4.135197639465332, + "step": 5655 + }, + { + "epoch": 78.5589519650655, + "grad_norm": 0.015059644356369972, + "learning_rate": 0.0006, + "loss": 4.255614280700684, + "step": 5656 + }, + { + "epoch": 78.57292576419214, + "grad_norm": 0.013833344914019108, + "learning_rate": 0.0006, + "loss": 4.301250457763672, + "step": 5657 + }, + { + "epoch": 78.58689956331878, + "grad_norm": 0.013665376231074333, + "learning_rate": 0.0006, + "loss": 4.1734232902526855, + "step": 5658 + }, + { + "epoch": 78.60087336244541, + "grad_norm": 0.01539086177945137, + "learning_rate": 0.0006, + "loss": 4.221459865570068, + "step": 5659 + }, + { + "epoch": 78.61484716157206, + "grad_norm": 0.013313917443156242, + "learning_rate": 0.0006, + "loss": 4.255619049072266, + "step": 5660 + }, + { + "epoch": 78.62882096069869, + "grad_norm": 0.01447741687297821, + "learning_rate": 0.0006, + "loss": 4.298219680786133, + "step": 5661 + }, + { + "epoch": 78.64279475982532, + "grad_norm": 0.012796915136277676, + "learning_rate": 0.0006, + "loss": 4.2542405128479, + "step": 5662 + }, + { + "epoch": 78.65676855895197, + "grad_norm": 0.014058141969144344, + "learning_rate": 0.0006, + "loss": 4.284477233886719, + "step": 5663 + }, + { + "epoch": 78.6707423580786, + "grad_norm": 0.012705069966614246, + "learning_rate": 0.0006, + "loss": 4.2826666831970215, + "step": 5664 + }, + { + "epoch": 78.68471615720524, + "grad_norm": 0.013245890848338604, + "learning_rate": 0.0006, + "loss": 4.371612548828125, + "step": 5665 + }, + { + "epoch": 78.69868995633188, + "grad_norm": 0.014192677102982998, + "learning_rate": 0.0006, + "loss": 4.267344951629639, + "step": 5666 + }, + { + "epoch": 78.71266375545852, + "grad_norm": 0.015346777625381947, + "learning_rate": 0.0006, + "loss": 4.222797393798828, + "step": 5667 + }, + { + "epoch": 78.72663755458515, + "grad_norm": 0.014068621210753918, + "learning_rate": 0.0006, + "loss": 4.172964572906494, + "step": 5668 + }, + { + "epoch": 78.74061135371178, + "grad_norm": 0.014364945702254772, + "learning_rate": 0.0006, + "loss": 4.259319305419922, + "step": 5669 + }, + { + "epoch": 78.75458515283843, + "grad_norm": 0.017337387427687645, + "learning_rate": 0.0006, + "loss": 4.29533576965332, + "step": 5670 + }, + { + "epoch": 78.76855895196506, + "grad_norm": 0.017416715621948242, + "learning_rate": 0.0006, + "loss": 4.340364456176758, + "step": 5671 + }, + { + "epoch": 78.78253275109171, + "grad_norm": 0.015925846993923187, + "learning_rate": 0.0006, + "loss": 4.177454948425293, + "step": 5672 + }, + { + "epoch": 78.79650655021834, + "grad_norm": 0.013130443170666695, + "learning_rate": 0.0006, + "loss": 4.277454853057861, + "step": 5673 + }, + { + "epoch": 78.81048034934499, + "grad_norm": 0.014628687873482704, + "learning_rate": 0.0006, + "loss": 4.273757457733154, + "step": 5674 + }, + { + "epoch": 78.82445414847162, + "grad_norm": 0.01601659320294857, + "learning_rate": 0.0006, + "loss": 4.281774520874023, + "step": 5675 + }, + { + "epoch": 78.83842794759825, + "grad_norm": 0.015443485230207443, + "learning_rate": 0.0006, + "loss": 4.199937343597412, + "step": 5676 + }, + { + "epoch": 78.8524017467249, + "grad_norm": 0.013934246264398098, + "learning_rate": 0.0006, + "loss": 4.331369876861572, + "step": 5677 + }, + { + "epoch": 78.86637554585153, + "grad_norm": 0.01572083681821823, + "learning_rate": 0.0006, + "loss": 4.2828474044799805, + "step": 5678 + }, + { + "epoch": 78.88034934497817, + "grad_norm": 0.013444120064377785, + "learning_rate": 0.0006, + "loss": 4.183859825134277, + "step": 5679 + }, + { + "epoch": 78.8943231441048, + "grad_norm": 0.01591254211962223, + "learning_rate": 0.0006, + "loss": 4.295256614685059, + "step": 5680 + }, + { + "epoch": 78.90829694323143, + "grad_norm": 0.017556684091687202, + "learning_rate": 0.0006, + "loss": 4.132272720336914, + "step": 5681 + }, + { + "epoch": 78.92227074235808, + "grad_norm": 0.018184322863817215, + "learning_rate": 0.0006, + "loss": 4.251728057861328, + "step": 5682 + }, + { + "epoch": 78.93624454148471, + "grad_norm": 0.016207285225391388, + "learning_rate": 0.0006, + "loss": 4.205544471740723, + "step": 5683 + }, + { + "epoch": 78.95021834061136, + "grad_norm": 0.014478376135230064, + "learning_rate": 0.0006, + "loss": 4.182462215423584, + "step": 5684 + }, + { + "epoch": 78.96419213973799, + "grad_norm": 0.014255235902965069, + "learning_rate": 0.0006, + "loss": 4.253366947174072, + "step": 5685 + }, + { + "epoch": 78.97816593886463, + "grad_norm": 0.01524396613240242, + "learning_rate": 0.0006, + "loss": 4.221786022186279, + "step": 5686 + }, + { + "epoch": 78.99213973799127, + "grad_norm": 0.015057693235576153, + "learning_rate": 0.0006, + "loss": 4.208454608917236, + "step": 5687 + }, + { + "epoch": 79.0, + "grad_norm": 0.015478022396564484, + "learning_rate": 0.0006, + "loss": 4.336383819580078, + "step": 5688 + }, + { + "epoch": 79.0, + "eval_loss": 4.641749382019043, + "eval_runtime": 56.8167, + "eval_samples_per_second": 42.98, + "eval_steps_per_second": 1.355, + "step": 5688 + }, + { + "epoch": 79.01397379912663, + "grad_norm": 0.014143316075205803, + "learning_rate": 0.0006, + "loss": 4.260974884033203, + "step": 5689 + }, + { + "epoch": 79.02794759825328, + "grad_norm": 0.013120883144438267, + "learning_rate": 0.0006, + "loss": 4.139886856079102, + "step": 5690 + }, + { + "epoch": 79.04192139737991, + "grad_norm": 0.015019661746919155, + "learning_rate": 0.0006, + "loss": 4.267004013061523, + "step": 5691 + }, + { + "epoch": 79.05589519650655, + "grad_norm": 0.014891284517943859, + "learning_rate": 0.0006, + "loss": 4.205019950866699, + "step": 5692 + }, + { + "epoch": 79.06986899563319, + "grad_norm": 0.015527728945016861, + "learning_rate": 0.0006, + "loss": 4.1513566970825195, + "step": 5693 + }, + { + "epoch": 79.08384279475983, + "grad_norm": 0.017527181655168533, + "learning_rate": 0.0006, + "loss": 4.148828983306885, + "step": 5694 + }, + { + "epoch": 79.09781659388646, + "grad_norm": 0.01632206328213215, + "learning_rate": 0.0006, + "loss": 4.195789813995361, + "step": 5695 + }, + { + "epoch": 79.1117903930131, + "grad_norm": 0.01590503193438053, + "learning_rate": 0.0006, + "loss": 4.199155330657959, + "step": 5696 + }, + { + "epoch": 79.12576419213974, + "grad_norm": 0.016489997506141663, + "learning_rate": 0.0006, + "loss": 4.155154705047607, + "step": 5697 + }, + { + "epoch": 79.13973799126637, + "grad_norm": 0.0161368977278471, + "learning_rate": 0.0006, + "loss": 4.164384841918945, + "step": 5698 + }, + { + "epoch": 79.15371179039302, + "grad_norm": 0.018574809655547142, + "learning_rate": 0.0006, + "loss": 4.1898193359375, + "step": 5699 + }, + { + "epoch": 79.16768558951965, + "grad_norm": 0.018987983465194702, + "learning_rate": 0.0006, + "loss": 4.219719409942627, + "step": 5700 + }, + { + "epoch": 79.18165938864628, + "grad_norm": 0.01668606884777546, + "learning_rate": 0.0006, + "loss": 4.152734756469727, + "step": 5701 + }, + { + "epoch": 79.19563318777293, + "grad_norm": 0.014046299271285534, + "learning_rate": 0.0006, + "loss": 4.152964115142822, + "step": 5702 + }, + { + "epoch": 79.20960698689956, + "grad_norm": 0.01628032699227333, + "learning_rate": 0.0006, + "loss": 4.299479961395264, + "step": 5703 + }, + { + "epoch": 79.2235807860262, + "grad_norm": 0.017998795956373215, + "learning_rate": 0.0006, + "loss": 4.231203079223633, + "step": 5704 + }, + { + "epoch": 79.23755458515284, + "grad_norm": 0.014799647964537144, + "learning_rate": 0.0006, + "loss": 4.3530192375183105, + "step": 5705 + }, + { + "epoch": 79.25152838427948, + "grad_norm": 0.014640678651630878, + "learning_rate": 0.0006, + "loss": 4.287322998046875, + "step": 5706 + }, + { + "epoch": 79.26550218340611, + "grad_norm": 0.0172616895288229, + "learning_rate": 0.0006, + "loss": 4.148009777069092, + "step": 5707 + }, + { + "epoch": 79.27947598253274, + "grad_norm": 0.01811790093779564, + "learning_rate": 0.0006, + "loss": 4.269433975219727, + "step": 5708 + }, + { + "epoch": 79.29344978165939, + "grad_norm": 0.01764000579714775, + "learning_rate": 0.0006, + "loss": 4.292638778686523, + "step": 5709 + }, + { + "epoch": 79.30742358078602, + "grad_norm": 0.018030935898423195, + "learning_rate": 0.0006, + "loss": 4.267889976501465, + "step": 5710 + }, + { + "epoch": 79.32139737991267, + "grad_norm": 0.017860891297459602, + "learning_rate": 0.0006, + "loss": 4.211110591888428, + "step": 5711 + }, + { + "epoch": 79.3353711790393, + "grad_norm": 0.017295073717832565, + "learning_rate": 0.0006, + "loss": 4.224645614624023, + "step": 5712 + }, + { + "epoch": 79.34934497816595, + "grad_norm": 0.015891101211309433, + "learning_rate": 0.0006, + "loss": 4.224797248840332, + "step": 5713 + }, + { + "epoch": 79.36331877729258, + "grad_norm": 0.014680417254567146, + "learning_rate": 0.0006, + "loss": 4.182242393493652, + "step": 5714 + }, + { + "epoch": 79.37729257641921, + "grad_norm": 0.016621140763163567, + "learning_rate": 0.0006, + "loss": 4.186334609985352, + "step": 5715 + }, + { + "epoch": 79.39126637554585, + "grad_norm": 0.016207249835133553, + "learning_rate": 0.0006, + "loss": 4.277074813842773, + "step": 5716 + }, + { + "epoch": 79.40524017467249, + "grad_norm": 0.015375478193163872, + "learning_rate": 0.0006, + "loss": 4.154317855834961, + "step": 5717 + }, + { + "epoch": 79.41921397379913, + "grad_norm": 0.013453228399157524, + "learning_rate": 0.0006, + "loss": 4.265270233154297, + "step": 5718 + }, + { + "epoch": 79.43318777292576, + "grad_norm": 0.014312694780528545, + "learning_rate": 0.0006, + "loss": 4.232220649719238, + "step": 5719 + }, + { + "epoch": 79.44716157205241, + "grad_norm": 0.014759554527699947, + "learning_rate": 0.0006, + "loss": 4.272275924682617, + "step": 5720 + }, + { + "epoch": 79.46113537117904, + "grad_norm": 0.015885505825281143, + "learning_rate": 0.0006, + "loss": 4.2229719161987305, + "step": 5721 + }, + { + "epoch": 79.47510917030567, + "grad_norm": 0.016516581177711487, + "learning_rate": 0.0006, + "loss": 4.200624465942383, + "step": 5722 + }, + { + "epoch": 79.48908296943232, + "grad_norm": 0.018254555761814117, + "learning_rate": 0.0006, + "loss": 4.19648551940918, + "step": 5723 + }, + { + "epoch": 79.50305676855895, + "grad_norm": 0.01467090006917715, + "learning_rate": 0.0006, + "loss": 4.135190486907959, + "step": 5724 + }, + { + "epoch": 79.5170305676856, + "grad_norm": 0.013474797829985619, + "learning_rate": 0.0006, + "loss": 4.311361312866211, + "step": 5725 + }, + { + "epoch": 79.53100436681223, + "grad_norm": 0.01742764748632908, + "learning_rate": 0.0006, + "loss": 4.1845903396606445, + "step": 5726 + }, + { + "epoch": 79.54497816593886, + "grad_norm": 0.018786882981657982, + "learning_rate": 0.0006, + "loss": 4.163942337036133, + "step": 5727 + }, + { + "epoch": 79.5589519650655, + "grad_norm": 0.017831219360232353, + "learning_rate": 0.0006, + "loss": 4.236685276031494, + "step": 5728 + }, + { + "epoch": 79.57292576419214, + "grad_norm": 0.016334986314177513, + "learning_rate": 0.0006, + "loss": 4.255853652954102, + "step": 5729 + }, + { + "epoch": 79.58689956331878, + "grad_norm": 0.0138487434014678, + "learning_rate": 0.0006, + "loss": 4.227874279022217, + "step": 5730 + }, + { + "epoch": 79.60087336244541, + "grad_norm": 0.01531760673969984, + "learning_rate": 0.0006, + "loss": 4.221432685852051, + "step": 5731 + }, + { + "epoch": 79.61484716157206, + "grad_norm": 0.01603361964225769, + "learning_rate": 0.0006, + "loss": 4.278326511383057, + "step": 5732 + }, + { + "epoch": 79.62882096069869, + "grad_norm": 0.013149895705282688, + "learning_rate": 0.0006, + "loss": 4.2474517822265625, + "step": 5733 + }, + { + "epoch": 79.64279475982532, + "grad_norm": 0.013428786769509315, + "learning_rate": 0.0006, + "loss": 4.1701178550720215, + "step": 5734 + }, + { + "epoch": 79.65676855895197, + "grad_norm": 0.014305585995316505, + "learning_rate": 0.0006, + "loss": 4.250778675079346, + "step": 5735 + }, + { + "epoch": 79.6707423580786, + "grad_norm": 0.014544566161930561, + "learning_rate": 0.0006, + "loss": 4.243236541748047, + "step": 5736 + }, + { + "epoch": 79.68471615720524, + "grad_norm": 0.013524414040148258, + "learning_rate": 0.0006, + "loss": 4.255457878112793, + "step": 5737 + }, + { + "epoch": 79.69868995633188, + "grad_norm": 0.014750641770660877, + "learning_rate": 0.0006, + "loss": 4.202816486358643, + "step": 5738 + }, + { + "epoch": 79.71266375545852, + "grad_norm": 0.015090773813426495, + "learning_rate": 0.0006, + "loss": 4.3053178787231445, + "step": 5739 + }, + { + "epoch": 79.72663755458515, + "grad_norm": 0.01606169529259205, + "learning_rate": 0.0006, + "loss": 4.224017143249512, + "step": 5740 + }, + { + "epoch": 79.74061135371178, + "grad_norm": 0.016303520649671555, + "learning_rate": 0.0006, + "loss": 4.190760612487793, + "step": 5741 + }, + { + "epoch": 79.75458515283843, + "grad_norm": 0.01465672068297863, + "learning_rate": 0.0006, + "loss": 4.24506950378418, + "step": 5742 + }, + { + "epoch": 79.76855895196506, + "grad_norm": 0.012616301886737347, + "learning_rate": 0.0006, + "loss": 4.174039840698242, + "step": 5743 + }, + { + "epoch": 79.78253275109171, + "grad_norm": 0.01358284242451191, + "learning_rate": 0.0006, + "loss": 4.236380577087402, + "step": 5744 + }, + { + "epoch": 79.79650655021834, + "grad_norm": 0.01557177770882845, + "learning_rate": 0.0006, + "loss": 4.102962493896484, + "step": 5745 + }, + { + "epoch": 79.81048034934499, + "grad_norm": 0.015584317035973072, + "learning_rate": 0.0006, + "loss": 4.242603302001953, + "step": 5746 + }, + { + "epoch": 79.82445414847162, + "grad_norm": 0.012597334571182728, + "learning_rate": 0.0006, + "loss": 4.202791690826416, + "step": 5747 + }, + { + "epoch": 79.83842794759825, + "grad_norm": 0.013722898438572884, + "learning_rate": 0.0006, + "loss": 4.204328536987305, + "step": 5748 + }, + { + "epoch": 79.8524017467249, + "grad_norm": 0.014669225551187992, + "learning_rate": 0.0006, + "loss": 4.164331912994385, + "step": 5749 + }, + { + "epoch": 79.86637554585153, + "grad_norm": 0.015012740157544613, + "learning_rate": 0.0006, + "loss": 4.260565280914307, + "step": 5750 + }, + { + "epoch": 79.88034934497817, + "grad_norm": 0.016026340425014496, + "learning_rate": 0.0006, + "loss": 4.299723148345947, + "step": 5751 + }, + { + "epoch": 79.8943231441048, + "grad_norm": 0.01626049168407917, + "learning_rate": 0.0006, + "loss": 4.207252502441406, + "step": 5752 + }, + { + "epoch": 79.90829694323143, + "grad_norm": 0.015415354631841183, + "learning_rate": 0.0006, + "loss": 4.151934623718262, + "step": 5753 + }, + { + "epoch": 79.92227074235808, + "grad_norm": 0.014158242382109165, + "learning_rate": 0.0006, + "loss": 4.3461995124816895, + "step": 5754 + }, + { + "epoch": 79.93624454148471, + "grad_norm": 0.01585628278553486, + "learning_rate": 0.0006, + "loss": 4.334761619567871, + "step": 5755 + }, + { + "epoch": 79.95021834061136, + "grad_norm": 0.013710759580135345, + "learning_rate": 0.0006, + "loss": 4.180401802062988, + "step": 5756 + }, + { + "epoch": 79.96419213973799, + "grad_norm": 0.014159204438328743, + "learning_rate": 0.0006, + "loss": 4.322331428527832, + "step": 5757 + }, + { + "epoch": 79.97816593886463, + "grad_norm": 0.015153195708990097, + "learning_rate": 0.0006, + "loss": 4.287478923797607, + "step": 5758 + }, + { + "epoch": 79.99213973799127, + "grad_norm": 0.014100227504968643, + "learning_rate": 0.0006, + "loss": 4.231807708740234, + "step": 5759 + }, + { + "epoch": 80.0, + "grad_norm": 0.015248682349920273, + "learning_rate": 0.0006, + "loss": 4.261106014251709, + "step": 5760 + }, + { + "epoch": 80.0, + "eval_loss": 4.610264301300049, + "eval_runtime": 56.514, + "eval_samples_per_second": 43.211, + "eval_steps_per_second": 1.362, + "step": 5760 + }, + { + "epoch": 80.01397379912663, + "grad_norm": 0.014730271883308887, + "learning_rate": 0.0006, + "loss": 4.108556747436523, + "step": 5761 + }, + { + "epoch": 80.02794759825328, + "grad_norm": 0.014575716108083725, + "learning_rate": 0.0006, + "loss": 4.169658660888672, + "step": 5762 + }, + { + "epoch": 80.04192139737991, + "grad_norm": 0.014706958085298538, + "learning_rate": 0.0006, + "loss": 4.252662658691406, + "step": 5763 + }, + { + "epoch": 80.05589519650655, + "grad_norm": 0.01561224739998579, + "learning_rate": 0.0006, + "loss": 4.193892955780029, + "step": 5764 + }, + { + "epoch": 80.06986899563319, + "grad_norm": 0.014702324755489826, + "learning_rate": 0.0006, + "loss": 4.1889238357543945, + "step": 5765 + }, + { + "epoch": 80.08384279475983, + "grad_norm": 0.015565741807222366, + "learning_rate": 0.0006, + "loss": 4.243549823760986, + "step": 5766 + }, + { + "epoch": 80.09781659388646, + "grad_norm": 0.01647181063890457, + "learning_rate": 0.0006, + "loss": 4.275446891784668, + "step": 5767 + }, + { + "epoch": 80.1117903930131, + "grad_norm": 0.018177716061472893, + "learning_rate": 0.0006, + "loss": 4.12040376663208, + "step": 5768 + }, + { + "epoch": 80.12576419213974, + "grad_norm": 0.01745191030204296, + "learning_rate": 0.0006, + "loss": 4.140371322631836, + "step": 5769 + }, + { + "epoch": 80.13973799126637, + "grad_norm": 0.01568993367254734, + "learning_rate": 0.0006, + "loss": 4.258280277252197, + "step": 5770 + }, + { + "epoch": 80.15371179039302, + "grad_norm": 0.01636597141623497, + "learning_rate": 0.0006, + "loss": 4.179052352905273, + "step": 5771 + }, + { + "epoch": 80.16768558951965, + "grad_norm": 0.016839100047945976, + "learning_rate": 0.0006, + "loss": 4.127636432647705, + "step": 5772 + }, + { + "epoch": 80.18165938864628, + "grad_norm": 0.015471016988158226, + "learning_rate": 0.0006, + "loss": 4.181325912475586, + "step": 5773 + }, + { + "epoch": 80.19563318777293, + "grad_norm": 0.015836089849472046, + "learning_rate": 0.0006, + "loss": 4.142168998718262, + "step": 5774 + }, + { + "epoch": 80.20960698689956, + "grad_norm": 0.014782954007387161, + "learning_rate": 0.0006, + "loss": 4.308799743652344, + "step": 5775 + }, + { + "epoch": 80.2235807860262, + "grad_norm": 0.014040893875062466, + "learning_rate": 0.0006, + "loss": 4.216936111450195, + "step": 5776 + }, + { + "epoch": 80.23755458515284, + "grad_norm": 0.01420280709862709, + "learning_rate": 0.0006, + "loss": 4.235230445861816, + "step": 5777 + }, + { + "epoch": 80.25152838427948, + "grad_norm": 0.013341007754206657, + "learning_rate": 0.0006, + "loss": 4.369109153747559, + "step": 5778 + }, + { + "epoch": 80.26550218340611, + "grad_norm": 0.014208327978849411, + "learning_rate": 0.0006, + "loss": 4.219112396240234, + "step": 5779 + }, + { + "epoch": 80.27947598253274, + "grad_norm": 0.013825052417814732, + "learning_rate": 0.0006, + "loss": 4.1983489990234375, + "step": 5780 + }, + { + "epoch": 80.29344978165939, + "grad_norm": 0.012916515581309795, + "learning_rate": 0.0006, + "loss": 4.204582691192627, + "step": 5781 + }, + { + "epoch": 80.30742358078602, + "grad_norm": 0.013071243651211262, + "learning_rate": 0.0006, + "loss": 4.17643928527832, + "step": 5782 + }, + { + "epoch": 80.32139737991267, + "grad_norm": 0.015507174655795097, + "learning_rate": 0.0006, + "loss": 4.172976016998291, + "step": 5783 + }, + { + "epoch": 80.3353711790393, + "grad_norm": 0.014967243187129498, + "learning_rate": 0.0006, + "loss": 4.141168594360352, + "step": 5784 + }, + { + "epoch": 80.34934497816595, + "grad_norm": 0.015629924833774567, + "learning_rate": 0.0006, + "loss": 4.253526210784912, + "step": 5785 + }, + { + "epoch": 80.36331877729258, + "grad_norm": 0.01607954502105713, + "learning_rate": 0.0006, + "loss": 4.249395370483398, + "step": 5786 + }, + { + "epoch": 80.37729257641921, + "grad_norm": 0.01500706560909748, + "learning_rate": 0.0006, + "loss": 4.279012680053711, + "step": 5787 + }, + { + "epoch": 80.39126637554585, + "grad_norm": 0.014343900606036186, + "learning_rate": 0.0006, + "loss": 4.268509864807129, + "step": 5788 + }, + { + "epoch": 80.40524017467249, + "grad_norm": 0.014682197012007236, + "learning_rate": 0.0006, + "loss": 4.206334114074707, + "step": 5789 + }, + { + "epoch": 80.41921397379913, + "grad_norm": 0.014989163726568222, + "learning_rate": 0.0006, + "loss": 4.143601417541504, + "step": 5790 + }, + { + "epoch": 80.43318777292576, + "grad_norm": 0.015001608058810234, + "learning_rate": 0.0006, + "loss": 4.164111137390137, + "step": 5791 + }, + { + "epoch": 80.44716157205241, + "grad_norm": 0.015222722664475441, + "learning_rate": 0.0006, + "loss": 4.2369914054870605, + "step": 5792 + }, + { + "epoch": 80.46113537117904, + "grad_norm": 0.014126183465123177, + "learning_rate": 0.0006, + "loss": 4.2209930419921875, + "step": 5793 + }, + { + "epoch": 80.47510917030567, + "grad_norm": 0.013065959326922894, + "learning_rate": 0.0006, + "loss": 4.231184959411621, + "step": 5794 + }, + { + "epoch": 80.48908296943232, + "grad_norm": 0.013833962380886078, + "learning_rate": 0.0006, + "loss": 4.226434707641602, + "step": 5795 + }, + { + "epoch": 80.50305676855895, + "grad_norm": 0.01387125626206398, + "learning_rate": 0.0006, + "loss": 4.241328716278076, + "step": 5796 + }, + { + "epoch": 80.5170305676856, + "grad_norm": 0.014145063236355782, + "learning_rate": 0.0006, + "loss": 4.2268805503845215, + "step": 5797 + }, + { + "epoch": 80.53100436681223, + "grad_norm": 0.012990240007638931, + "learning_rate": 0.0006, + "loss": 4.1718339920043945, + "step": 5798 + }, + { + "epoch": 80.54497816593886, + "grad_norm": 0.014185411855578423, + "learning_rate": 0.0006, + "loss": 4.232397556304932, + "step": 5799 + }, + { + "epoch": 80.5589519650655, + "grad_norm": 0.014260523952543736, + "learning_rate": 0.0006, + "loss": 4.282734394073486, + "step": 5800 + }, + { + "epoch": 80.57292576419214, + "grad_norm": 0.012542120181024075, + "learning_rate": 0.0006, + "loss": 4.134551525115967, + "step": 5801 + }, + { + "epoch": 80.58689956331878, + "grad_norm": 0.013078657910227776, + "learning_rate": 0.0006, + "loss": 4.229815483093262, + "step": 5802 + }, + { + "epoch": 80.60087336244541, + "grad_norm": 0.012821056880056858, + "learning_rate": 0.0006, + "loss": 4.208278179168701, + "step": 5803 + }, + { + "epoch": 80.61484716157206, + "grad_norm": 0.012964180670678616, + "learning_rate": 0.0006, + "loss": 4.188477039337158, + "step": 5804 + }, + { + "epoch": 80.62882096069869, + "grad_norm": 0.01631746254861355, + "learning_rate": 0.0006, + "loss": 4.3214497566223145, + "step": 5805 + }, + { + "epoch": 80.64279475982532, + "grad_norm": 0.01726640947163105, + "learning_rate": 0.0006, + "loss": 4.279359817504883, + "step": 5806 + }, + { + "epoch": 80.65676855895197, + "grad_norm": 0.018778258934617043, + "learning_rate": 0.0006, + "loss": 4.294047832489014, + "step": 5807 + }, + { + "epoch": 80.6707423580786, + "grad_norm": 0.01796831749379635, + "learning_rate": 0.0006, + "loss": 4.314162254333496, + "step": 5808 + }, + { + "epoch": 80.68471615720524, + "grad_norm": 0.016495538875460625, + "learning_rate": 0.0006, + "loss": 4.273871421813965, + "step": 5809 + }, + { + "epoch": 80.69868995633188, + "grad_norm": 0.013590306043624878, + "learning_rate": 0.0006, + "loss": 4.14961051940918, + "step": 5810 + }, + { + "epoch": 80.71266375545852, + "grad_norm": 0.014670426957309246, + "learning_rate": 0.0006, + "loss": 4.211853981018066, + "step": 5811 + }, + { + "epoch": 80.72663755458515, + "grad_norm": 0.01641341857612133, + "learning_rate": 0.0006, + "loss": 4.211948871612549, + "step": 5812 + }, + { + "epoch": 80.74061135371178, + "grad_norm": 0.016365470364689827, + "learning_rate": 0.0006, + "loss": 4.200774192810059, + "step": 5813 + }, + { + "epoch": 80.75458515283843, + "grad_norm": 0.01613970287144184, + "learning_rate": 0.0006, + "loss": 4.288322448730469, + "step": 5814 + }, + { + "epoch": 80.76855895196506, + "grad_norm": 0.016742777079343796, + "learning_rate": 0.0006, + "loss": 4.23801326751709, + "step": 5815 + }, + { + "epoch": 80.78253275109171, + "grad_norm": 0.015351069159805775, + "learning_rate": 0.0006, + "loss": 4.263832092285156, + "step": 5816 + }, + { + "epoch": 80.79650655021834, + "grad_norm": 0.013435336761176586, + "learning_rate": 0.0006, + "loss": 4.160857200622559, + "step": 5817 + }, + { + "epoch": 80.81048034934499, + "grad_norm": 0.01474886853247881, + "learning_rate": 0.0006, + "loss": 4.216025352478027, + "step": 5818 + }, + { + "epoch": 80.82445414847162, + "grad_norm": 0.01586935855448246, + "learning_rate": 0.0006, + "loss": 4.256402969360352, + "step": 5819 + }, + { + "epoch": 80.83842794759825, + "grad_norm": 0.014596572145819664, + "learning_rate": 0.0006, + "loss": 4.142584800720215, + "step": 5820 + }, + { + "epoch": 80.8524017467249, + "grad_norm": 0.01501593366265297, + "learning_rate": 0.0006, + "loss": 4.250462055206299, + "step": 5821 + }, + { + "epoch": 80.86637554585153, + "grad_norm": 0.014773511327803135, + "learning_rate": 0.0006, + "loss": 4.253305912017822, + "step": 5822 + }, + { + "epoch": 80.88034934497817, + "grad_norm": 0.013246326707303524, + "learning_rate": 0.0006, + "loss": 4.170881271362305, + "step": 5823 + }, + { + "epoch": 80.8943231441048, + "grad_norm": 0.013013756833970547, + "learning_rate": 0.0006, + "loss": 4.264721870422363, + "step": 5824 + }, + { + "epoch": 80.90829694323143, + "grad_norm": 0.013155903667211533, + "learning_rate": 0.0006, + "loss": 4.274667739868164, + "step": 5825 + }, + { + "epoch": 80.92227074235808, + "grad_norm": 0.01448952779173851, + "learning_rate": 0.0006, + "loss": 4.24777889251709, + "step": 5826 + }, + { + "epoch": 80.93624454148471, + "grad_norm": 0.01415231078863144, + "learning_rate": 0.0006, + "loss": 4.22014856338501, + "step": 5827 + }, + { + "epoch": 80.95021834061136, + "grad_norm": 0.014796273782849312, + "learning_rate": 0.0006, + "loss": 4.22036600112915, + "step": 5828 + }, + { + "epoch": 80.96419213973799, + "grad_norm": 0.013005326502025127, + "learning_rate": 0.0006, + "loss": 4.276123046875, + "step": 5829 + }, + { + "epoch": 80.97816593886463, + "grad_norm": 0.012851177714765072, + "learning_rate": 0.0006, + "loss": 4.248575687408447, + "step": 5830 + }, + { + "epoch": 80.99213973799127, + "grad_norm": 0.013740002177655697, + "learning_rate": 0.0006, + "loss": 4.051140308380127, + "step": 5831 + }, + { + "epoch": 81.0, + "grad_norm": 0.015296104364097118, + "learning_rate": 0.0006, + "loss": 4.251155853271484, + "step": 5832 + }, + { + "epoch": 81.0, + "eval_loss": 4.594127655029297, + "eval_runtime": 56.0959, + "eval_samples_per_second": 43.533, + "eval_steps_per_second": 1.373, + "step": 5832 + }, + { + "epoch": 81.01397379912663, + "grad_norm": 0.015804558992385864, + "learning_rate": 0.0006, + "loss": 4.152894020080566, + "step": 5833 + }, + { + "epoch": 81.02794759825328, + "grad_norm": 0.017000854015350342, + "learning_rate": 0.0006, + "loss": 4.266067981719971, + "step": 5834 + }, + { + "epoch": 81.04192139737991, + "grad_norm": 0.016351008787751198, + "learning_rate": 0.0006, + "loss": 4.192703723907471, + "step": 5835 + }, + { + "epoch": 81.05589519650655, + "grad_norm": 0.014703895896673203, + "learning_rate": 0.0006, + "loss": 4.23554801940918, + "step": 5836 + }, + { + "epoch": 81.06986899563319, + "grad_norm": 0.01619596965610981, + "learning_rate": 0.0006, + "loss": 4.233000755310059, + "step": 5837 + }, + { + "epoch": 81.08384279475983, + "grad_norm": 0.01815493032336235, + "learning_rate": 0.0006, + "loss": 4.200490951538086, + "step": 5838 + }, + { + "epoch": 81.09781659388646, + "grad_norm": 0.018305106088519096, + "learning_rate": 0.0006, + "loss": 4.154242515563965, + "step": 5839 + }, + { + "epoch": 81.1117903930131, + "grad_norm": 0.0200856514275074, + "learning_rate": 0.0006, + "loss": 4.25820779800415, + "step": 5840 + }, + { + "epoch": 81.12576419213974, + "grad_norm": 0.02101440541446209, + "learning_rate": 0.0006, + "loss": 4.168209075927734, + "step": 5841 + }, + { + "epoch": 81.13973799126637, + "grad_norm": 0.022829681634902954, + "learning_rate": 0.0006, + "loss": 4.25560998916626, + "step": 5842 + }, + { + "epoch": 81.15371179039302, + "grad_norm": 0.0195082426071167, + "learning_rate": 0.0006, + "loss": 4.2035627365112305, + "step": 5843 + }, + { + "epoch": 81.16768558951965, + "grad_norm": 0.01716248318552971, + "learning_rate": 0.0006, + "loss": 4.11322021484375, + "step": 5844 + }, + { + "epoch": 81.18165938864628, + "grad_norm": 0.02034664712846279, + "learning_rate": 0.0006, + "loss": 4.179389953613281, + "step": 5845 + }, + { + "epoch": 81.19563318777293, + "grad_norm": 0.023578710854053497, + "learning_rate": 0.0006, + "loss": 4.265695571899414, + "step": 5846 + }, + { + "epoch": 81.20960698689956, + "grad_norm": 0.019244182854890823, + "learning_rate": 0.0006, + "loss": 4.292043209075928, + "step": 5847 + }, + { + "epoch": 81.2235807860262, + "grad_norm": 0.01577291637659073, + "learning_rate": 0.0006, + "loss": 4.250776290893555, + "step": 5848 + }, + { + "epoch": 81.23755458515284, + "grad_norm": 0.01617090217769146, + "learning_rate": 0.0006, + "loss": 4.193098068237305, + "step": 5849 + }, + { + "epoch": 81.25152838427948, + "grad_norm": 0.016090184450149536, + "learning_rate": 0.0006, + "loss": 4.112357139587402, + "step": 5850 + }, + { + "epoch": 81.26550218340611, + "grad_norm": 0.01441365759819746, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 5851 + }, + { + "epoch": 81.27947598253274, + "grad_norm": 0.014122162945568562, + "learning_rate": 0.0006, + "loss": 4.276175022125244, + "step": 5852 + }, + { + "epoch": 81.29344978165939, + "grad_norm": 0.014859487302601337, + "learning_rate": 0.0006, + "loss": 4.094699859619141, + "step": 5853 + }, + { + "epoch": 81.30742358078602, + "grad_norm": 0.013624267652630806, + "learning_rate": 0.0006, + "loss": 4.2648773193359375, + "step": 5854 + }, + { + "epoch": 81.32139737991267, + "grad_norm": 0.01417913381010294, + "learning_rate": 0.0006, + "loss": 4.228567123413086, + "step": 5855 + }, + { + "epoch": 81.3353711790393, + "grad_norm": 0.014560805633664131, + "learning_rate": 0.0006, + "loss": 4.148930549621582, + "step": 5856 + }, + { + "epoch": 81.34934497816595, + "grad_norm": 0.013662154786288738, + "learning_rate": 0.0006, + "loss": 4.192107677459717, + "step": 5857 + }, + { + "epoch": 81.36331877729258, + "grad_norm": 0.013434980995953083, + "learning_rate": 0.0006, + "loss": 4.162052154541016, + "step": 5858 + }, + { + "epoch": 81.37729257641921, + "grad_norm": 0.01300742756575346, + "learning_rate": 0.0006, + "loss": 4.262385368347168, + "step": 5859 + }, + { + "epoch": 81.39126637554585, + "grad_norm": 0.012544351629912853, + "learning_rate": 0.0006, + "loss": 4.215371131896973, + "step": 5860 + }, + { + "epoch": 81.40524017467249, + "grad_norm": 0.013663645833730698, + "learning_rate": 0.0006, + "loss": 4.247312545776367, + "step": 5861 + }, + { + "epoch": 81.41921397379913, + "grad_norm": 0.013127296231687069, + "learning_rate": 0.0006, + "loss": 4.087685585021973, + "step": 5862 + }, + { + "epoch": 81.43318777292576, + "grad_norm": 0.014081938192248344, + "learning_rate": 0.0006, + "loss": 4.272648811340332, + "step": 5863 + }, + { + "epoch": 81.44716157205241, + "grad_norm": 0.015470641665160656, + "learning_rate": 0.0006, + "loss": 4.278578758239746, + "step": 5864 + }, + { + "epoch": 81.46113537117904, + "grad_norm": 0.016524532809853554, + "learning_rate": 0.0006, + "loss": 4.118268013000488, + "step": 5865 + }, + { + "epoch": 81.47510917030567, + "grad_norm": 0.016242722049355507, + "learning_rate": 0.0006, + "loss": 4.182405471801758, + "step": 5866 + }, + { + "epoch": 81.48908296943232, + "grad_norm": 0.01557017583400011, + "learning_rate": 0.0006, + "loss": 4.229862213134766, + "step": 5867 + }, + { + "epoch": 81.50305676855895, + "grad_norm": 0.01562870666384697, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 5868 + }, + { + "epoch": 81.5170305676856, + "grad_norm": 0.018606074154376984, + "learning_rate": 0.0006, + "loss": 4.148681640625, + "step": 5869 + }, + { + "epoch": 81.53100436681223, + "grad_norm": 0.017740730196237564, + "learning_rate": 0.0006, + "loss": 4.321932792663574, + "step": 5870 + }, + { + "epoch": 81.54497816593886, + "grad_norm": 0.014517677947878838, + "learning_rate": 0.0006, + "loss": 4.22862434387207, + "step": 5871 + }, + { + "epoch": 81.5589519650655, + "grad_norm": 0.01389714702963829, + "learning_rate": 0.0006, + "loss": 4.198000907897949, + "step": 5872 + }, + { + "epoch": 81.57292576419214, + "grad_norm": 0.015348682180047035, + "learning_rate": 0.0006, + "loss": 4.293485164642334, + "step": 5873 + }, + { + "epoch": 81.58689956331878, + "grad_norm": 0.015965791419148445, + "learning_rate": 0.0006, + "loss": 4.23481559753418, + "step": 5874 + }, + { + "epoch": 81.60087336244541, + "grad_norm": 0.015662726014852524, + "learning_rate": 0.0006, + "loss": 4.158148765563965, + "step": 5875 + }, + { + "epoch": 81.61484716157206, + "grad_norm": 0.014402080327272415, + "learning_rate": 0.0006, + "loss": 4.132723808288574, + "step": 5876 + }, + { + "epoch": 81.62882096069869, + "grad_norm": 0.016846125945448875, + "learning_rate": 0.0006, + "loss": 4.1654462814331055, + "step": 5877 + }, + { + "epoch": 81.64279475982532, + "grad_norm": 0.0196097232401371, + "learning_rate": 0.0006, + "loss": 4.228450298309326, + "step": 5878 + }, + { + "epoch": 81.65676855895197, + "grad_norm": 0.016005294397473335, + "learning_rate": 0.0006, + "loss": 4.230222225189209, + "step": 5879 + }, + { + "epoch": 81.6707423580786, + "grad_norm": 0.014685059897601604, + "learning_rate": 0.0006, + "loss": 4.1367387771606445, + "step": 5880 + }, + { + "epoch": 81.68471615720524, + "grad_norm": 0.01647144928574562, + "learning_rate": 0.0006, + "loss": 4.096675395965576, + "step": 5881 + }, + { + "epoch": 81.69868995633188, + "grad_norm": 0.015978610143065453, + "learning_rate": 0.0006, + "loss": 4.232292175292969, + "step": 5882 + }, + { + "epoch": 81.71266375545852, + "grad_norm": 0.01510288193821907, + "learning_rate": 0.0006, + "loss": 4.274591445922852, + "step": 5883 + }, + { + "epoch": 81.72663755458515, + "grad_norm": 0.014942476525902748, + "learning_rate": 0.0006, + "loss": 4.161482810974121, + "step": 5884 + }, + { + "epoch": 81.74061135371178, + "grad_norm": 0.014595299027860165, + "learning_rate": 0.0006, + "loss": 4.204291343688965, + "step": 5885 + }, + { + "epoch": 81.75458515283843, + "grad_norm": 0.014952429570257664, + "learning_rate": 0.0006, + "loss": 4.242499351501465, + "step": 5886 + }, + { + "epoch": 81.76855895196506, + "grad_norm": 0.014761554077267647, + "learning_rate": 0.0006, + "loss": 4.20393180847168, + "step": 5887 + }, + { + "epoch": 81.78253275109171, + "grad_norm": 0.014749647118151188, + "learning_rate": 0.0006, + "loss": 4.227565765380859, + "step": 5888 + }, + { + "epoch": 81.79650655021834, + "grad_norm": 0.014384493231773376, + "learning_rate": 0.0006, + "loss": 4.282228946685791, + "step": 5889 + }, + { + "epoch": 81.81048034934499, + "grad_norm": 0.01260330155491829, + "learning_rate": 0.0006, + "loss": 4.2882866859436035, + "step": 5890 + }, + { + "epoch": 81.82445414847162, + "grad_norm": 0.014055311679840088, + "learning_rate": 0.0006, + "loss": 4.22351598739624, + "step": 5891 + }, + { + "epoch": 81.83842794759825, + "grad_norm": 0.013674065470695496, + "learning_rate": 0.0006, + "loss": 4.240716934204102, + "step": 5892 + }, + { + "epoch": 81.8524017467249, + "grad_norm": 0.01225196197628975, + "learning_rate": 0.0006, + "loss": 4.2958598136901855, + "step": 5893 + }, + { + "epoch": 81.86637554585153, + "grad_norm": 0.013538680970668793, + "learning_rate": 0.0006, + "loss": 4.24342679977417, + "step": 5894 + }, + { + "epoch": 81.88034934497817, + "grad_norm": 0.013151394203305244, + "learning_rate": 0.0006, + "loss": 4.081999778747559, + "step": 5895 + }, + { + "epoch": 81.8943231441048, + "grad_norm": 0.014260678552091122, + "learning_rate": 0.0006, + "loss": 4.156423568725586, + "step": 5896 + }, + { + "epoch": 81.90829694323143, + "grad_norm": 0.01329776830971241, + "learning_rate": 0.0006, + "loss": 4.1713995933532715, + "step": 5897 + }, + { + "epoch": 81.92227074235808, + "grad_norm": 0.014623276889324188, + "learning_rate": 0.0006, + "loss": 4.244891166687012, + "step": 5898 + }, + { + "epoch": 81.93624454148471, + "grad_norm": 0.015760324895381927, + "learning_rate": 0.0006, + "loss": 4.076528549194336, + "step": 5899 + }, + { + "epoch": 81.95021834061136, + "grad_norm": 0.014331766404211521, + "learning_rate": 0.0006, + "loss": 4.283037185668945, + "step": 5900 + }, + { + "epoch": 81.96419213973799, + "grad_norm": 0.012383842840790749, + "learning_rate": 0.0006, + "loss": 4.117308616638184, + "step": 5901 + }, + { + "epoch": 81.97816593886463, + "grad_norm": 0.01250660140067339, + "learning_rate": 0.0006, + "loss": 4.241250991821289, + "step": 5902 + }, + { + "epoch": 81.99213973799127, + "grad_norm": 0.012276900932192802, + "learning_rate": 0.0006, + "loss": 4.141091823577881, + "step": 5903 + }, + { + "epoch": 82.0, + "grad_norm": 0.01588737778365612, + "learning_rate": 0.0006, + "loss": 4.065532684326172, + "step": 5904 + }, + { + "epoch": 82.0, + "eval_loss": 4.612304210662842, + "eval_runtime": 57.5411, + "eval_samples_per_second": 42.439, + "eval_steps_per_second": 1.338, + "step": 5904 + }, + { + "epoch": 82.01397379912663, + "grad_norm": 0.017433544620871544, + "learning_rate": 0.0006, + "loss": 4.202497482299805, + "step": 5905 + }, + { + "epoch": 82.02794759825328, + "grad_norm": 0.017260415479540825, + "learning_rate": 0.0006, + "loss": 4.099949359893799, + "step": 5906 + }, + { + "epoch": 82.04192139737991, + "grad_norm": 0.01577908545732498, + "learning_rate": 0.0006, + "loss": 4.17806339263916, + "step": 5907 + }, + { + "epoch": 82.05589519650655, + "grad_norm": 0.016093818470835686, + "learning_rate": 0.0006, + "loss": 4.219491004943848, + "step": 5908 + }, + { + "epoch": 82.06986899563319, + "grad_norm": 0.015426302328705788, + "learning_rate": 0.0006, + "loss": 4.207982063293457, + "step": 5909 + }, + { + "epoch": 82.08384279475983, + "grad_norm": 0.01605442352592945, + "learning_rate": 0.0006, + "loss": 4.265942573547363, + "step": 5910 + }, + { + "epoch": 82.09781659388646, + "grad_norm": 0.01787225715816021, + "learning_rate": 0.0006, + "loss": 4.151535511016846, + "step": 5911 + }, + { + "epoch": 82.1117903930131, + "grad_norm": 0.01750164106488228, + "learning_rate": 0.0006, + "loss": 4.133655071258545, + "step": 5912 + }, + { + "epoch": 82.12576419213974, + "grad_norm": 0.01602957211434841, + "learning_rate": 0.0006, + "loss": 4.176972389221191, + "step": 5913 + }, + { + "epoch": 82.13973799126637, + "grad_norm": 0.01553812064230442, + "learning_rate": 0.0006, + "loss": 4.125692367553711, + "step": 5914 + }, + { + "epoch": 82.15371179039302, + "grad_norm": 0.013978508301079273, + "learning_rate": 0.0006, + "loss": 4.207271575927734, + "step": 5915 + }, + { + "epoch": 82.16768558951965, + "grad_norm": 0.013561118394136429, + "learning_rate": 0.0006, + "loss": 4.238930702209473, + "step": 5916 + }, + { + "epoch": 82.18165938864628, + "grad_norm": 0.013285507448017597, + "learning_rate": 0.0006, + "loss": 4.129464626312256, + "step": 5917 + }, + { + "epoch": 82.19563318777293, + "grad_norm": 0.013903839513659477, + "learning_rate": 0.0006, + "loss": 4.214109420776367, + "step": 5918 + }, + { + "epoch": 82.20960698689956, + "grad_norm": 0.014426837675273418, + "learning_rate": 0.0006, + "loss": 4.143196105957031, + "step": 5919 + }, + { + "epoch": 82.2235807860262, + "grad_norm": 0.015347165986895561, + "learning_rate": 0.0006, + "loss": 4.235439300537109, + "step": 5920 + }, + { + "epoch": 82.23755458515284, + "grad_norm": 0.014266039244830608, + "learning_rate": 0.0006, + "loss": 3.991617202758789, + "step": 5921 + }, + { + "epoch": 82.25152838427948, + "grad_norm": 0.013101593591272831, + "learning_rate": 0.0006, + "loss": 4.266017913818359, + "step": 5922 + }, + { + "epoch": 82.26550218340611, + "grad_norm": 0.013741881586611271, + "learning_rate": 0.0006, + "loss": 4.224879741668701, + "step": 5923 + }, + { + "epoch": 82.27947598253274, + "grad_norm": 0.014206945896148682, + "learning_rate": 0.0006, + "loss": 4.102272987365723, + "step": 5924 + }, + { + "epoch": 82.29344978165939, + "grad_norm": 0.013798723928630352, + "learning_rate": 0.0006, + "loss": 4.30418586730957, + "step": 5925 + }, + { + "epoch": 82.30742358078602, + "grad_norm": 0.015470122918486595, + "learning_rate": 0.0006, + "loss": 4.2513322830200195, + "step": 5926 + }, + { + "epoch": 82.32139737991267, + "grad_norm": 0.01655442826449871, + "learning_rate": 0.0006, + "loss": 4.223002910614014, + "step": 5927 + }, + { + "epoch": 82.3353711790393, + "grad_norm": 0.017170004546642303, + "learning_rate": 0.0006, + "loss": 4.253904819488525, + "step": 5928 + }, + { + "epoch": 82.34934497816595, + "grad_norm": 0.017595812678337097, + "learning_rate": 0.0006, + "loss": 4.184927940368652, + "step": 5929 + }, + { + "epoch": 82.36331877729258, + "grad_norm": 0.015073256567120552, + "learning_rate": 0.0006, + "loss": 4.202239036560059, + "step": 5930 + }, + { + "epoch": 82.37729257641921, + "grad_norm": 0.014310941100120544, + "learning_rate": 0.0006, + "loss": 4.124934673309326, + "step": 5931 + }, + { + "epoch": 82.39126637554585, + "grad_norm": 0.013671821914613247, + "learning_rate": 0.0006, + "loss": 4.138919353485107, + "step": 5932 + }, + { + "epoch": 82.40524017467249, + "grad_norm": 0.014103314839303493, + "learning_rate": 0.0006, + "loss": 4.211289405822754, + "step": 5933 + }, + { + "epoch": 82.41921397379913, + "grad_norm": 0.014272668398916721, + "learning_rate": 0.0006, + "loss": 4.06290340423584, + "step": 5934 + }, + { + "epoch": 82.43318777292576, + "grad_norm": 0.01387016475200653, + "learning_rate": 0.0006, + "loss": 4.101334571838379, + "step": 5935 + }, + { + "epoch": 82.44716157205241, + "grad_norm": 0.013318683952093124, + "learning_rate": 0.0006, + "loss": 4.206409931182861, + "step": 5936 + }, + { + "epoch": 82.46113537117904, + "grad_norm": 0.013781268149614334, + "learning_rate": 0.0006, + "loss": 4.189474105834961, + "step": 5937 + }, + { + "epoch": 82.47510917030567, + "grad_norm": 0.013195851817727089, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 5938 + }, + { + "epoch": 82.48908296943232, + "grad_norm": 0.014358163811266422, + "learning_rate": 0.0006, + "loss": 4.203262805938721, + "step": 5939 + }, + { + "epoch": 82.50305676855895, + "grad_norm": 0.01403636671602726, + "learning_rate": 0.0006, + "loss": 4.079258441925049, + "step": 5940 + }, + { + "epoch": 82.5170305676856, + "grad_norm": 0.013272535055875778, + "learning_rate": 0.0006, + "loss": 4.19197940826416, + "step": 5941 + }, + { + "epoch": 82.53100436681223, + "grad_norm": 0.013787264935672283, + "learning_rate": 0.0006, + "loss": 4.12995719909668, + "step": 5942 + }, + { + "epoch": 82.54497816593886, + "grad_norm": 0.015020057559013367, + "learning_rate": 0.0006, + "loss": 4.148890495300293, + "step": 5943 + }, + { + "epoch": 82.5589519650655, + "grad_norm": 0.014199157245457172, + "learning_rate": 0.0006, + "loss": 4.201516151428223, + "step": 5944 + }, + { + "epoch": 82.57292576419214, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.240894317626953, + "step": 5945 + }, + { + "epoch": 82.58689956331878, + "grad_norm": 0.014767560176551342, + "learning_rate": 0.0006, + "loss": 4.173320770263672, + "step": 5946 + }, + { + "epoch": 82.60087336244541, + "grad_norm": 0.01612078957259655, + "learning_rate": 0.0006, + "loss": 4.266448020935059, + "step": 5947 + }, + { + "epoch": 82.61484716157206, + "grad_norm": 0.016706543043255806, + "learning_rate": 0.0006, + "loss": 4.229398250579834, + "step": 5948 + }, + { + "epoch": 82.62882096069869, + "grad_norm": 0.015630796551704407, + "learning_rate": 0.0006, + "loss": 4.181282997131348, + "step": 5949 + }, + { + "epoch": 82.64279475982532, + "grad_norm": 0.013103844597935677, + "learning_rate": 0.0006, + "loss": 4.309617042541504, + "step": 5950 + }, + { + "epoch": 82.65676855895197, + "grad_norm": 0.016136541962623596, + "learning_rate": 0.0006, + "loss": 4.224159240722656, + "step": 5951 + }, + { + "epoch": 82.6707423580786, + "grad_norm": 0.016160814091563225, + "learning_rate": 0.0006, + "loss": 4.267970085144043, + "step": 5952 + }, + { + "epoch": 82.68471615720524, + "grad_norm": 0.01672680303454399, + "learning_rate": 0.0006, + "loss": 4.230506896972656, + "step": 5953 + }, + { + "epoch": 82.69868995633188, + "grad_norm": 0.0159586351364851, + "learning_rate": 0.0006, + "loss": 4.159542083740234, + "step": 5954 + }, + { + "epoch": 82.71266375545852, + "grad_norm": 0.01461219135671854, + "learning_rate": 0.0006, + "loss": 4.27863073348999, + "step": 5955 + }, + { + "epoch": 82.72663755458515, + "grad_norm": 0.015088613145053387, + "learning_rate": 0.0006, + "loss": 4.272418975830078, + "step": 5956 + }, + { + "epoch": 82.74061135371178, + "grad_norm": 0.01594499684870243, + "learning_rate": 0.0006, + "loss": 4.271326065063477, + "step": 5957 + }, + { + "epoch": 82.75458515283843, + "grad_norm": 0.016648447141051292, + "learning_rate": 0.0006, + "loss": 4.141478538513184, + "step": 5958 + }, + { + "epoch": 82.76855895196506, + "grad_norm": 0.01540316641330719, + "learning_rate": 0.0006, + "loss": 4.111814498901367, + "step": 5959 + }, + { + "epoch": 82.78253275109171, + "grad_norm": 0.01634364202618599, + "learning_rate": 0.0006, + "loss": 4.3113112449646, + "step": 5960 + }, + { + "epoch": 82.79650655021834, + "grad_norm": 0.014768977649509907, + "learning_rate": 0.0006, + "loss": 4.184217929840088, + "step": 5961 + }, + { + "epoch": 82.81048034934499, + "grad_norm": 0.013453755527734756, + "learning_rate": 0.0006, + "loss": 4.247729301452637, + "step": 5962 + }, + { + "epoch": 82.82445414847162, + "grad_norm": 0.014731907285749912, + "learning_rate": 0.0006, + "loss": 4.352839469909668, + "step": 5963 + }, + { + "epoch": 82.83842794759825, + "grad_norm": 0.016047345474362373, + "learning_rate": 0.0006, + "loss": 4.257283687591553, + "step": 5964 + }, + { + "epoch": 82.8524017467249, + "grad_norm": 0.017141170799732208, + "learning_rate": 0.0006, + "loss": 4.22850227355957, + "step": 5965 + }, + { + "epoch": 82.86637554585153, + "grad_norm": 0.01370396837592125, + "learning_rate": 0.0006, + "loss": 4.1169257164001465, + "step": 5966 + }, + { + "epoch": 82.88034934497817, + "grad_norm": 0.012060081586241722, + "learning_rate": 0.0006, + "loss": 4.195242404937744, + "step": 5967 + }, + { + "epoch": 82.8943231441048, + "grad_norm": 0.015528378076851368, + "learning_rate": 0.0006, + "loss": 4.294939994812012, + "step": 5968 + }, + { + "epoch": 82.90829694323143, + "grad_norm": 0.018151041120290756, + "learning_rate": 0.0006, + "loss": 4.137192726135254, + "step": 5969 + }, + { + "epoch": 82.92227074235808, + "grad_norm": 0.017895735800266266, + "learning_rate": 0.0006, + "loss": 4.174036026000977, + "step": 5970 + }, + { + "epoch": 82.93624454148471, + "grad_norm": 0.015610386617481709, + "learning_rate": 0.0006, + "loss": 4.230136871337891, + "step": 5971 + }, + { + "epoch": 82.95021834061136, + "grad_norm": 0.01339928898960352, + "learning_rate": 0.0006, + "loss": 4.199074745178223, + "step": 5972 + }, + { + "epoch": 82.96419213973799, + "grad_norm": 0.015454214997589588, + "learning_rate": 0.0006, + "loss": 4.216109752655029, + "step": 5973 + }, + { + "epoch": 82.97816593886463, + "grad_norm": 0.014978132210671902, + "learning_rate": 0.0006, + "loss": 4.131511688232422, + "step": 5974 + }, + { + "epoch": 82.99213973799127, + "grad_norm": 0.015611122362315655, + "learning_rate": 0.0006, + "loss": 4.246518135070801, + "step": 5975 + }, + { + "epoch": 83.0, + "grad_norm": 0.016682764515280724, + "learning_rate": 0.0006, + "loss": 4.110680103302002, + "step": 5976 + }, + { + "epoch": 83.0, + "eval_loss": 4.641707897186279, + "eval_runtime": 57.2799, + "eval_samples_per_second": 42.633, + "eval_steps_per_second": 1.344, + "step": 5976 + }, + { + "epoch": 83.01397379912663, + "grad_norm": 0.015007982030510902, + "learning_rate": 0.0006, + "loss": 4.100912094116211, + "step": 5977 + }, + { + "epoch": 83.02794759825328, + "grad_norm": 0.017609436064958572, + "learning_rate": 0.0006, + "loss": 4.233399868011475, + "step": 5978 + }, + { + "epoch": 83.04192139737991, + "grad_norm": 0.019817272201180458, + "learning_rate": 0.0006, + "loss": 4.191173076629639, + "step": 5979 + }, + { + "epoch": 83.05589519650655, + "grad_norm": 0.017752140760421753, + "learning_rate": 0.0006, + "loss": 4.044437885284424, + "step": 5980 + }, + { + "epoch": 83.06986899563319, + "grad_norm": 0.015482760034501553, + "learning_rate": 0.0006, + "loss": 4.319931983947754, + "step": 5981 + }, + { + "epoch": 83.08384279475983, + "grad_norm": 0.015153988264501095, + "learning_rate": 0.0006, + "loss": 4.095277309417725, + "step": 5982 + }, + { + "epoch": 83.09781659388646, + "grad_norm": 0.017025282606482506, + "learning_rate": 0.0006, + "loss": 4.235110759735107, + "step": 5983 + }, + { + "epoch": 83.1117903930131, + "grad_norm": 0.01566586270928383, + "learning_rate": 0.0006, + "loss": 4.1934685707092285, + "step": 5984 + }, + { + "epoch": 83.12576419213974, + "grad_norm": 0.016419263556599617, + "learning_rate": 0.0006, + "loss": 4.234419822692871, + "step": 5985 + }, + { + "epoch": 83.13973799126637, + "grad_norm": 0.017132675275206566, + "learning_rate": 0.0006, + "loss": 4.221447944641113, + "step": 5986 + }, + { + "epoch": 83.15371179039302, + "grad_norm": 0.016138629987835884, + "learning_rate": 0.0006, + "loss": 4.199830055236816, + "step": 5987 + }, + { + "epoch": 83.16768558951965, + "grad_norm": 0.016210174188017845, + "learning_rate": 0.0006, + "loss": 4.081717491149902, + "step": 5988 + }, + { + "epoch": 83.18165938864628, + "grad_norm": 0.016055308282375336, + "learning_rate": 0.0006, + "loss": 4.247655391693115, + "step": 5989 + }, + { + "epoch": 83.19563318777293, + "grad_norm": 0.016132952645421028, + "learning_rate": 0.0006, + "loss": 4.262535572052002, + "step": 5990 + }, + { + "epoch": 83.20960698689956, + "grad_norm": 0.01779935695230961, + "learning_rate": 0.0006, + "loss": 4.153120040893555, + "step": 5991 + }, + { + "epoch": 83.2235807860262, + "grad_norm": 0.018360646441578865, + "learning_rate": 0.0006, + "loss": 4.232644081115723, + "step": 5992 + }, + { + "epoch": 83.23755458515284, + "grad_norm": 0.019629675894975662, + "learning_rate": 0.0006, + "loss": 3.9758734703063965, + "step": 5993 + }, + { + "epoch": 83.25152838427948, + "grad_norm": 0.017819996923208237, + "learning_rate": 0.0006, + "loss": 4.196932315826416, + "step": 5994 + }, + { + "epoch": 83.26550218340611, + "grad_norm": 0.01557681243866682, + "learning_rate": 0.0006, + "loss": 4.122654914855957, + "step": 5995 + }, + { + "epoch": 83.27947598253274, + "grad_norm": 0.017858853563666344, + "learning_rate": 0.0006, + "loss": 4.185330867767334, + "step": 5996 + }, + { + "epoch": 83.29344978165939, + "grad_norm": 0.018065953627228737, + "learning_rate": 0.0006, + "loss": 4.218471050262451, + "step": 5997 + }, + { + "epoch": 83.30742358078602, + "grad_norm": 0.017247939482331276, + "learning_rate": 0.0006, + "loss": 4.127076148986816, + "step": 5998 + }, + { + "epoch": 83.32139737991267, + "grad_norm": 0.017047952860593796, + "learning_rate": 0.0006, + "loss": 4.131356716156006, + "step": 5999 + }, + { + "epoch": 83.3353711790393, + "grad_norm": 0.015905629843473434, + "learning_rate": 0.0006, + "loss": 4.2239789962768555, + "step": 6000 + }, + { + "epoch": 83.34934497816595, + "grad_norm": 0.016522347927093506, + "learning_rate": 0.0006, + "loss": 4.054897785186768, + "step": 6001 + }, + { + "epoch": 83.36331877729258, + "grad_norm": 0.01622571051120758, + "learning_rate": 0.0006, + "loss": 4.284210681915283, + "step": 6002 + }, + { + "epoch": 83.37729257641921, + "grad_norm": 0.01640467159450054, + "learning_rate": 0.0006, + "loss": 4.268840789794922, + "step": 6003 + }, + { + "epoch": 83.39126637554585, + "grad_norm": 0.015698859468102455, + "learning_rate": 0.0006, + "loss": 4.179793357849121, + "step": 6004 + }, + { + "epoch": 83.40524017467249, + "grad_norm": 0.016694528982043266, + "learning_rate": 0.0006, + "loss": 4.132297515869141, + "step": 6005 + }, + { + "epoch": 83.41921397379913, + "grad_norm": 0.015345900319516659, + "learning_rate": 0.0006, + "loss": 4.090163230895996, + "step": 6006 + }, + { + "epoch": 83.43318777292576, + "grad_norm": 0.014722758904099464, + "learning_rate": 0.0006, + "loss": 4.179166793823242, + "step": 6007 + }, + { + "epoch": 83.44716157205241, + "grad_norm": 0.014884887263178825, + "learning_rate": 0.0006, + "loss": 4.137659072875977, + "step": 6008 + }, + { + "epoch": 83.46113537117904, + "grad_norm": 0.014041902497410774, + "learning_rate": 0.0006, + "loss": 4.1202073097229, + "step": 6009 + }, + { + "epoch": 83.47510917030567, + "grad_norm": 0.012799333781003952, + "learning_rate": 0.0006, + "loss": 4.092607498168945, + "step": 6010 + }, + { + "epoch": 83.48908296943232, + "grad_norm": 0.013272707350552082, + "learning_rate": 0.0006, + "loss": 4.132715225219727, + "step": 6011 + }, + { + "epoch": 83.50305676855895, + "grad_norm": 0.012934371829032898, + "learning_rate": 0.0006, + "loss": 4.230278968811035, + "step": 6012 + }, + { + "epoch": 83.5170305676856, + "grad_norm": 0.0120006762444973, + "learning_rate": 0.0006, + "loss": 4.140068531036377, + "step": 6013 + }, + { + "epoch": 83.53100436681223, + "grad_norm": 0.013616700656712055, + "learning_rate": 0.0006, + "loss": 4.20107364654541, + "step": 6014 + }, + { + "epoch": 83.54497816593886, + "grad_norm": 0.013957851566374302, + "learning_rate": 0.0006, + "loss": 4.124945163726807, + "step": 6015 + }, + { + "epoch": 83.5589519650655, + "grad_norm": 0.012762055732309818, + "learning_rate": 0.0006, + "loss": 4.022052764892578, + "step": 6016 + }, + { + "epoch": 83.57292576419214, + "grad_norm": 0.013080576434731483, + "learning_rate": 0.0006, + "loss": 4.103252410888672, + "step": 6017 + }, + { + "epoch": 83.58689956331878, + "grad_norm": 0.013957500457763672, + "learning_rate": 0.0006, + "loss": 4.235089302062988, + "step": 6018 + }, + { + "epoch": 83.60087336244541, + "grad_norm": 0.014707164838910103, + "learning_rate": 0.0006, + "loss": 4.191570281982422, + "step": 6019 + }, + { + "epoch": 83.61484716157206, + "grad_norm": 0.012807869352400303, + "learning_rate": 0.0006, + "loss": 4.294543266296387, + "step": 6020 + }, + { + "epoch": 83.62882096069869, + "grad_norm": 0.012027994729578495, + "learning_rate": 0.0006, + "loss": 4.13849401473999, + "step": 6021 + }, + { + "epoch": 83.64279475982532, + "grad_norm": 0.012147724628448486, + "learning_rate": 0.0006, + "loss": 4.127481937408447, + "step": 6022 + }, + { + "epoch": 83.65676855895197, + "grad_norm": 0.013241392560303211, + "learning_rate": 0.0006, + "loss": 4.218972682952881, + "step": 6023 + }, + { + "epoch": 83.6707423580786, + "grad_norm": 0.014071648940443993, + "learning_rate": 0.0006, + "loss": 4.096724033355713, + "step": 6024 + }, + { + "epoch": 83.68471615720524, + "grad_norm": 0.015212797559797764, + "learning_rate": 0.0006, + "loss": 4.27097225189209, + "step": 6025 + }, + { + "epoch": 83.69868995633188, + "grad_norm": 0.017647385597229004, + "learning_rate": 0.0006, + "loss": 4.209835529327393, + "step": 6026 + }, + { + "epoch": 83.71266375545852, + "grad_norm": 0.01709647662937641, + "learning_rate": 0.0006, + "loss": 4.042106628417969, + "step": 6027 + }, + { + "epoch": 83.72663755458515, + "grad_norm": 0.014530927874147892, + "learning_rate": 0.0006, + "loss": 4.181731700897217, + "step": 6028 + }, + { + "epoch": 83.74061135371178, + "grad_norm": 0.015356711111962795, + "learning_rate": 0.0006, + "loss": 4.314183235168457, + "step": 6029 + }, + { + "epoch": 83.75458515283843, + "grad_norm": 0.015500029549002647, + "learning_rate": 0.0006, + "loss": 4.06776762008667, + "step": 6030 + }, + { + "epoch": 83.76855895196506, + "grad_norm": 0.015351332724094391, + "learning_rate": 0.0006, + "loss": 4.185141563415527, + "step": 6031 + }, + { + "epoch": 83.78253275109171, + "grad_norm": 0.016678154468536377, + "learning_rate": 0.0006, + "loss": 4.179529666900635, + "step": 6032 + }, + { + "epoch": 83.79650655021834, + "grad_norm": 0.0156265702098608, + "learning_rate": 0.0006, + "loss": 4.238584041595459, + "step": 6033 + }, + { + "epoch": 83.81048034934499, + "grad_norm": 0.016418596729636192, + "learning_rate": 0.0006, + "loss": 4.1593523025512695, + "step": 6034 + }, + { + "epoch": 83.82445414847162, + "grad_norm": 0.016113586723804474, + "learning_rate": 0.0006, + "loss": 4.157470226287842, + "step": 6035 + }, + { + "epoch": 83.83842794759825, + "grad_norm": 0.014434042386710644, + "learning_rate": 0.0006, + "loss": 4.196226119995117, + "step": 6036 + }, + { + "epoch": 83.8524017467249, + "grad_norm": 0.012647042982280254, + "learning_rate": 0.0006, + "loss": 4.130589485168457, + "step": 6037 + }, + { + "epoch": 83.86637554585153, + "grad_norm": 0.014412877149879932, + "learning_rate": 0.0006, + "loss": 4.154845237731934, + "step": 6038 + }, + { + "epoch": 83.88034934497817, + "grad_norm": 0.014222027733922005, + "learning_rate": 0.0006, + "loss": 4.189724922180176, + "step": 6039 + }, + { + "epoch": 83.8943231441048, + "grad_norm": 0.013084255158901215, + "learning_rate": 0.0006, + "loss": 4.303338527679443, + "step": 6040 + }, + { + "epoch": 83.90829694323143, + "grad_norm": 0.012659232132136822, + "learning_rate": 0.0006, + "loss": 4.08192253112793, + "step": 6041 + }, + { + "epoch": 83.92227074235808, + "grad_norm": 0.01246276218444109, + "learning_rate": 0.0006, + "loss": 4.283324718475342, + "step": 6042 + }, + { + "epoch": 83.93624454148471, + "grad_norm": 0.01340118795633316, + "learning_rate": 0.0006, + "loss": 4.117908477783203, + "step": 6043 + }, + { + "epoch": 83.95021834061136, + "grad_norm": 0.014496722258627415, + "learning_rate": 0.0006, + "loss": 4.110318660736084, + "step": 6044 + }, + { + "epoch": 83.96419213973799, + "grad_norm": 0.013348298147320747, + "learning_rate": 0.0006, + "loss": 4.185139179229736, + "step": 6045 + }, + { + "epoch": 83.97816593886463, + "grad_norm": 0.013294277712702751, + "learning_rate": 0.0006, + "loss": 4.271645545959473, + "step": 6046 + }, + { + "epoch": 83.99213973799127, + "grad_norm": 0.014431594870984554, + "learning_rate": 0.0006, + "loss": 4.161383152008057, + "step": 6047 + }, + { + "epoch": 84.0, + "grad_norm": 0.016268927603960037, + "learning_rate": 0.0006, + "loss": 4.178324222564697, + "step": 6048 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.578475223519068e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6048/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6048/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6048/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6336/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6336/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a2ea71b3237632ffc28130e245c13a34ee03a134 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bcfdc4bd44a4657ef33e702c02341b87b4a18ccaac135c4d8c6ff5ff8e3e5e3 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6336/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..15000b58fd2fa4486c6968e8a39b11756d6a5e15 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ed1db5379aaee2d0aeeaea8df19fe785ad63c518130165e8d0d5613ea85bf89 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..9bfd8fd41317ce97ec5426ef00a336bb43093975 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5669c00dd9e0c288cf3502acb77b3f67f83fa66af28dd688459b2b61b04597d5 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..166c8af41498ffc28a7641d3b21dd7a6b15c17b7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e1e292f0fe3c7fb524391e6995752fb45102689b83575bb8bf380e55677c963 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..3b04d3b8510e2cb426c14bf493785c01f9b3953a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62c7b95ac7104656fd3a5a8a50a70e41f051ca80db9ee60b624c946b282b9466 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..b7ea885e549895088ffc7c40f579739debda147d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64fde3e7147a7a919872654aaa3b6ff513c5a49eeb3b65f8fdeea1def70174ae +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6336/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..806ff31f0128f007291ee83361dda9fba5eeb32b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79947657782b5ee38d1d49b023c1983388f9e08fc318792128b0f35b4e555422 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6336/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..7cceab1151587e723d3d114ba68783317d29dbd1 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/trainer_state.json @@ -0,0 +1,45082 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 88.0, + "eval_steps": 500, + "global_step": 6336, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + }, + { + "epoch": 76.0, + "eval_loss": 4.689640998840332, + "eval_runtime": 56.9095, + "eval_samples_per_second": 42.91, + "eval_steps_per_second": 1.353, + "step": 5472 + }, + { + "epoch": 76.01397379912663, + "grad_norm": 0.015351055189967155, + "learning_rate": 0.0006, + "loss": 4.348268985748291, + "step": 5473 + }, + { + "epoch": 76.02794759825328, + "grad_norm": 0.017164217308163643, + "learning_rate": 0.0006, + "loss": 4.244204998016357, + "step": 5474 + }, + { + "epoch": 76.04192139737991, + "grad_norm": 0.021624932065606117, + "learning_rate": 0.0006, + "loss": 4.311168670654297, + "step": 5475 + }, + { + "epoch": 76.05589519650655, + "grad_norm": 0.028712468221783638, + "learning_rate": 0.0006, + "loss": 4.322287082672119, + "step": 5476 + }, + { + "epoch": 76.06986899563319, + "grad_norm": 0.027915487065911293, + "learning_rate": 0.0006, + "loss": 4.381207466125488, + "step": 5477 + }, + { + "epoch": 76.08384279475983, + "grad_norm": 0.017955303192138672, + "learning_rate": 0.0006, + "loss": 4.2635016441345215, + "step": 5478 + }, + { + "epoch": 76.09781659388646, + "grad_norm": 0.019113514572381973, + "learning_rate": 0.0006, + "loss": 4.399471282958984, + "step": 5479 + }, + { + "epoch": 76.1117903930131, + "grad_norm": 0.017502542585134506, + "learning_rate": 0.0006, + "loss": 4.321204662322998, + "step": 5480 + }, + { + "epoch": 76.12576419213974, + "grad_norm": 0.01852525770664215, + "learning_rate": 0.0006, + "loss": 4.217884063720703, + "step": 5481 + }, + { + "epoch": 76.13973799126637, + "grad_norm": 0.019836828112602234, + "learning_rate": 0.0006, + "loss": 4.278532028198242, + "step": 5482 + }, + { + "epoch": 76.15371179039302, + "grad_norm": 0.02101975679397583, + "learning_rate": 0.0006, + "loss": 4.246720314025879, + "step": 5483 + }, + { + "epoch": 76.16768558951965, + "grad_norm": 0.021202990785241127, + "learning_rate": 0.0006, + "loss": 4.517517566680908, + "step": 5484 + }, + { + "epoch": 76.18165938864628, + "grad_norm": 0.021428769454360008, + "learning_rate": 0.0006, + "loss": 4.3373517990112305, + "step": 5485 + }, + { + "epoch": 76.19563318777293, + "grad_norm": 0.018263190984725952, + "learning_rate": 0.0006, + "loss": 4.402856826782227, + "step": 5486 + }, + { + "epoch": 76.20960698689956, + "grad_norm": 0.019025536254048347, + "learning_rate": 0.0006, + "loss": 4.314227104187012, + "step": 5487 + }, + { + "epoch": 76.2235807860262, + "grad_norm": 0.020918630063533783, + "learning_rate": 0.0006, + "loss": 4.454369068145752, + "step": 5488 + }, + { + "epoch": 76.23755458515284, + "grad_norm": 0.019307617098093033, + "learning_rate": 0.0006, + "loss": 4.247509002685547, + "step": 5489 + }, + { + "epoch": 76.25152838427948, + "grad_norm": 0.014956331811845303, + "learning_rate": 0.0006, + "loss": 4.365175247192383, + "step": 5490 + }, + { + "epoch": 76.26550218340611, + "grad_norm": 0.017213059589266777, + "learning_rate": 0.0006, + "loss": 4.384005069732666, + "step": 5491 + }, + { + "epoch": 76.27947598253274, + "grad_norm": 0.017575597390532494, + "learning_rate": 0.0006, + "loss": 4.41934871673584, + "step": 5492 + }, + { + "epoch": 76.29344978165939, + "grad_norm": 0.015879256650805473, + "learning_rate": 0.0006, + "loss": 4.26675271987915, + "step": 5493 + }, + { + "epoch": 76.30742358078602, + "grad_norm": 0.01564522460103035, + "learning_rate": 0.0006, + "loss": 4.277339935302734, + "step": 5494 + }, + { + "epoch": 76.32139737991267, + "grad_norm": 0.016414586454629898, + "learning_rate": 0.0006, + "loss": 4.4138593673706055, + "step": 5495 + }, + { + "epoch": 76.3353711790393, + "grad_norm": 0.014732337556779385, + "learning_rate": 0.0006, + "loss": 4.281006336212158, + "step": 5496 + }, + { + "epoch": 76.34934497816595, + "grad_norm": 0.015461455099284649, + "learning_rate": 0.0006, + "loss": 4.263422966003418, + "step": 5497 + }, + { + "epoch": 76.36331877729258, + "grad_norm": 0.014529814012348652, + "learning_rate": 0.0006, + "loss": 4.36165189743042, + "step": 5498 + }, + { + "epoch": 76.37729257641921, + "grad_norm": 0.01469483319669962, + "learning_rate": 0.0006, + "loss": 4.309642791748047, + "step": 5499 + }, + { + "epoch": 76.39126637554585, + "grad_norm": 0.016490574926137924, + "learning_rate": 0.0006, + "loss": 4.288150787353516, + "step": 5500 + }, + { + "epoch": 76.40524017467249, + "grad_norm": 0.014593565836548805, + "learning_rate": 0.0006, + "loss": 4.271614074707031, + "step": 5501 + }, + { + "epoch": 76.41921397379913, + "grad_norm": 0.012796535156667233, + "learning_rate": 0.0006, + "loss": 4.420042037963867, + "step": 5502 + }, + { + "epoch": 76.43318777292576, + "grad_norm": 0.016086025163531303, + "learning_rate": 0.0006, + "loss": 4.289943695068359, + "step": 5503 + }, + { + "epoch": 76.44716157205241, + "grad_norm": 0.015439470298588276, + "learning_rate": 0.0006, + "loss": 4.3023271560668945, + "step": 5504 + }, + { + "epoch": 76.46113537117904, + "grad_norm": 0.015066367574036121, + "learning_rate": 0.0006, + "loss": 4.197081565856934, + "step": 5505 + }, + { + "epoch": 76.47510917030567, + "grad_norm": 0.016653263941407204, + "learning_rate": 0.0006, + "loss": 4.333785057067871, + "step": 5506 + }, + { + "epoch": 76.48908296943232, + "grad_norm": 0.014791973866522312, + "learning_rate": 0.0006, + "loss": 4.323811054229736, + "step": 5507 + }, + { + "epoch": 76.50305676855895, + "grad_norm": 0.014694827608764172, + "learning_rate": 0.0006, + "loss": 4.195852279663086, + "step": 5508 + }, + { + "epoch": 76.5170305676856, + "grad_norm": 0.015120396390557289, + "learning_rate": 0.0006, + "loss": 4.42379903793335, + "step": 5509 + }, + { + "epoch": 76.53100436681223, + "grad_norm": 0.014584025368094444, + "learning_rate": 0.0006, + "loss": 4.2702531814575195, + "step": 5510 + }, + { + "epoch": 76.54497816593886, + "grad_norm": 0.013981613330543041, + "learning_rate": 0.0006, + "loss": 4.383029937744141, + "step": 5511 + }, + { + "epoch": 76.5589519650655, + "grad_norm": 0.012930642813444138, + "learning_rate": 0.0006, + "loss": 4.296136856079102, + "step": 5512 + }, + { + "epoch": 76.57292576419214, + "grad_norm": 0.013060989789664745, + "learning_rate": 0.0006, + "loss": 4.346380233764648, + "step": 5513 + }, + { + "epoch": 76.58689956331878, + "grad_norm": 0.013209798373281956, + "learning_rate": 0.0006, + "loss": 4.316067218780518, + "step": 5514 + }, + { + "epoch": 76.60087336244541, + "grad_norm": 0.012423216365277767, + "learning_rate": 0.0006, + "loss": 4.370274543762207, + "step": 5515 + }, + { + "epoch": 76.61484716157206, + "grad_norm": 0.013205411843955517, + "learning_rate": 0.0006, + "loss": 4.429969310760498, + "step": 5516 + }, + { + "epoch": 76.62882096069869, + "grad_norm": 0.013049827888607979, + "learning_rate": 0.0006, + "loss": 4.378549098968506, + "step": 5517 + }, + { + "epoch": 76.64279475982532, + "grad_norm": 0.012075880542397499, + "learning_rate": 0.0006, + "loss": 4.348293304443359, + "step": 5518 + }, + { + "epoch": 76.65676855895197, + "grad_norm": 0.014932847581803799, + "learning_rate": 0.0006, + "loss": 4.282538890838623, + "step": 5519 + }, + { + "epoch": 76.6707423580786, + "grad_norm": 0.015615998767316341, + "learning_rate": 0.0006, + "loss": 4.274641036987305, + "step": 5520 + }, + { + "epoch": 76.68471615720524, + "grad_norm": 0.01424513477832079, + "learning_rate": 0.0006, + "loss": 4.243166923522949, + "step": 5521 + }, + { + "epoch": 76.69868995633188, + "grad_norm": 0.012775546871125698, + "learning_rate": 0.0006, + "loss": 4.335611820220947, + "step": 5522 + }, + { + "epoch": 76.71266375545852, + "grad_norm": 0.014461303129792213, + "learning_rate": 0.0006, + "loss": 4.246976375579834, + "step": 5523 + }, + { + "epoch": 76.72663755458515, + "grad_norm": 0.0145436255261302, + "learning_rate": 0.0006, + "loss": 4.275895118713379, + "step": 5524 + }, + { + "epoch": 76.74061135371178, + "grad_norm": 0.014055085368454456, + "learning_rate": 0.0006, + "loss": 4.331108093261719, + "step": 5525 + }, + { + "epoch": 76.75458515283843, + "grad_norm": 0.011933894827961922, + "learning_rate": 0.0006, + "loss": 4.293721675872803, + "step": 5526 + }, + { + "epoch": 76.76855895196506, + "grad_norm": 0.01277392078191042, + "learning_rate": 0.0006, + "loss": 4.312374114990234, + "step": 5527 + }, + { + "epoch": 76.78253275109171, + "grad_norm": 0.014227936044335365, + "learning_rate": 0.0006, + "loss": 4.238729000091553, + "step": 5528 + }, + { + "epoch": 76.79650655021834, + "grad_norm": 0.01488024927675724, + "learning_rate": 0.0006, + "loss": 4.24587345123291, + "step": 5529 + }, + { + "epoch": 76.81048034934499, + "grad_norm": 0.015341250225901604, + "learning_rate": 0.0006, + "loss": 4.257378578186035, + "step": 5530 + }, + { + "epoch": 76.82445414847162, + "grad_norm": 0.015179437585175037, + "learning_rate": 0.0006, + "loss": 4.2812910079956055, + "step": 5531 + }, + { + "epoch": 76.83842794759825, + "grad_norm": 0.012814724817872047, + "learning_rate": 0.0006, + "loss": 4.2309465408325195, + "step": 5532 + }, + { + "epoch": 76.8524017467249, + "grad_norm": 0.013810716569423676, + "learning_rate": 0.0006, + "loss": 4.325292110443115, + "step": 5533 + }, + { + "epoch": 76.86637554585153, + "grad_norm": 0.017468232661485672, + "learning_rate": 0.0006, + "loss": 4.243788719177246, + "step": 5534 + }, + { + "epoch": 76.88034934497817, + "grad_norm": 0.01739857904613018, + "learning_rate": 0.0006, + "loss": 4.345390796661377, + "step": 5535 + }, + { + "epoch": 76.8943231441048, + "grad_norm": 0.015329061076045036, + "learning_rate": 0.0006, + "loss": 4.415142059326172, + "step": 5536 + }, + { + "epoch": 76.90829694323143, + "grad_norm": 0.01598426327109337, + "learning_rate": 0.0006, + "loss": 4.299558639526367, + "step": 5537 + }, + { + "epoch": 76.92227074235808, + "grad_norm": 0.01621454581618309, + "learning_rate": 0.0006, + "loss": 4.284354209899902, + "step": 5538 + }, + { + "epoch": 76.93624454148471, + "grad_norm": 0.015956968069076538, + "learning_rate": 0.0006, + "loss": 4.293837547302246, + "step": 5539 + }, + { + "epoch": 76.95021834061136, + "grad_norm": 0.01776493526995182, + "learning_rate": 0.0006, + "loss": 4.229729175567627, + "step": 5540 + }, + { + "epoch": 76.96419213973799, + "grad_norm": 0.015317477285861969, + "learning_rate": 0.0006, + "loss": 4.3328447341918945, + "step": 5541 + }, + { + "epoch": 76.97816593886463, + "grad_norm": 0.015140251256525517, + "learning_rate": 0.0006, + "loss": 4.310840606689453, + "step": 5542 + }, + { + "epoch": 76.99213973799127, + "grad_norm": 0.015326812863349915, + "learning_rate": 0.0006, + "loss": 4.230832099914551, + "step": 5543 + }, + { + "epoch": 77.0, + "grad_norm": 0.01607763208448887, + "learning_rate": 0.0006, + "loss": 4.2236328125, + "step": 5544 + }, + { + "epoch": 77.0, + "eval_loss": 4.662221431732178, + "eval_runtime": 56.5194, + "eval_samples_per_second": 43.206, + "eval_steps_per_second": 1.362, + "step": 5544 + }, + { + "epoch": 77.01397379912663, + "grad_norm": 0.014839510433375835, + "learning_rate": 0.0006, + "loss": 4.254111289978027, + "step": 5545 + }, + { + "epoch": 77.02794759825328, + "grad_norm": 0.01519741676747799, + "learning_rate": 0.0006, + "loss": 4.1607184410095215, + "step": 5546 + }, + { + "epoch": 77.04192139737991, + "grad_norm": 0.015472603030502796, + "learning_rate": 0.0006, + "loss": 4.323320388793945, + "step": 5547 + }, + { + "epoch": 77.05589519650655, + "grad_norm": 0.017937948927283287, + "learning_rate": 0.0006, + "loss": 4.308490753173828, + "step": 5548 + }, + { + "epoch": 77.06986899563319, + "grad_norm": 0.020890841260552406, + "learning_rate": 0.0006, + "loss": 4.259101867675781, + "step": 5549 + }, + { + "epoch": 77.08384279475983, + "grad_norm": 0.01892857998609543, + "learning_rate": 0.0006, + "loss": 4.386970520019531, + "step": 5550 + }, + { + "epoch": 77.09781659388646, + "grad_norm": 0.016583630815148354, + "learning_rate": 0.0006, + "loss": 4.233604907989502, + "step": 5551 + }, + { + "epoch": 77.1117903930131, + "grad_norm": 0.016063468530774117, + "learning_rate": 0.0006, + "loss": 4.339544296264648, + "step": 5552 + }, + { + "epoch": 77.12576419213974, + "grad_norm": 0.018326779827475548, + "learning_rate": 0.0006, + "loss": 4.367204666137695, + "step": 5553 + }, + { + "epoch": 77.13973799126637, + "grad_norm": 0.021590933203697205, + "learning_rate": 0.0006, + "loss": 4.352685928344727, + "step": 5554 + }, + { + "epoch": 77.15371179039302, + "grad_norm": 0.019643642008304596, + "learning_rate": 0.0006, + "loss": 4.258535385131836, + "step": 5555 + }, + { + "epoch": 77.16768558951965, + "grad_norm": 0.018476149067282677, + "learning_rate": 0.0006, + "loss": 4.232568740844727, + "step": 5556 + }, + { + "epoch": 77.18165938864628, + "grad_norm": 0.01584405079483986, + "learning_rate": 0.0006, + "loss": 4.258371353149414, + "step": 5557 + }, + { + "epoch": 77.19563318777293, + "grad_norm": 0.01575201191008091, + "learning_rate": 0.0006, + "loss": 4.183218002319336, + "step": 5558 + }, + { + "epoch": 77.20960698689956, + "grad_norm": 0.016871824860572815, + "learning_rate": 0.0006, + "loss": 4.194563865661621, + "step": 5559 + }, + { + "epoch": 77.2235807860262, + "grad_norm": 0.02059333398938179, + "learning_rate": 0.0006, + "loss": 4.195281982421875, + "step": 5560 + }, + { + "epoch": 77.23755458515284, + "grad_norm": 0.022640453651547432, + "learning_rate": 0.0006, + "loss": 4.279615879058838, + "step": 5561 + }, + { + "epoch": 77.25152838427948, + "grad_norm": 0.020335959270596504, + "learning_rate": 0.0006, + "loss": 4.25040864944458, + "step": 5562 + }, + { + "epoch": 77.26550218340611, + "grad_norm": 0.020010793581604958, + "learning_rate": 0.0006, + "loss": 4.291116714477539, + "step": 5563 + }, + { + "epoch": 77.27947598253274, + "grad_norm": 0.02109045907855034, + "learning_rate": 0.0006, + "loss": 4.364088535308838, + "step": 5564 + }, + { + "epoch": 77.29344978165939, + "grad_norm": 0.019519446417689323, + "learning_rate": 0.0006, + "loss": 4.329166412353516, + "step": 5565 + }, + { + "epoch": 77.30742358078602, + "grad_norm": 0.01689804159104824, + "learning_rate": 0.0006, + "loss": 4.244176864624023, + "step": 5566 + }, + { + "epoch": 77.32139737991267, + "grad_norm": 0.01893681474030018, + "learning_rate": 0.0006, + "loss": 4.239068984985352, + "step": 5567 + }, + { + "epoch": 77.3353711790393, + "grad_norm": 0.015998193994164467, + "learning_rate": 0.0006, + "loss": 4.173398017883301, + "step": 5568 + }, + { + "epoch": 77.34934497816595, + "grad_norm": 0.013892280869185925, + "learning_rate": 0.0006, + "loss": 4.287654876708984, + "step": 5569 + }, + { + "epoch": 77.36331877729258, + "grad_norm": 0.017060643061995506, + "learning_rate": 0.0006, + "loss": 4.398447036743164, + "step": 5570 + }, + { + "epoch": 77.37729257641921, + "grad_norm": 0.015168849378824234, + "learning_rate": 0.0006, + "loss": 4.15665864944458, + "step": 5571 + }, + { + "epoch": 77.39126637554585, + "grad_norm": 0.014886523596942425, + "learning_rate": 0.0006, + "loss": 4.196673393249512, + "step": 5572 + }, + { + "epoch": 77.40524017467249, + "grad_norm": 0.016941796988248825, + "learning_rate": 0.0006, + "loss": 4.314562797546387, + "step": 5573 + }, + { + "epoch": 77.41921397379913, + "grad_norm": 0.015353812836110592, + "learning_rate": 0.0006, + "loss": 4.409579277038574, + "step": 5574 + }, + { + "epoch": 77.43318777292576, + "grad_norm": 0.015192852355539799, + "learning_rate": 0.0006, + "loss": 4.392045974731445, + "step": 5575 + }, + { + "epoch": 77.44716157205241, + "grad_norm": 0.014232569374144077, + "learning_rate": 0.0006, + "loss": 4.21639347076416, + "step": 5576 + }, + { + "epoch": 77.46113537117904, + "grad_norm": 0.014453226700425148, + "learning_rate": 0.0006, + "loss": 4.2802276611328125, + "step": 5577 + }, + { + "epoch": 77.47510917030567, + "grad_norm": 0.012881236150860786, + "learning_rate": 0.0006, + "loss": 4.375980854034424, + "step": 5578 + }, + { + "epoch": 77.48908296943232, + "grad_norm": 0.01583976298570633, + "learning_rate": 0.0006, + "loss": 4.264520645141602, + "step": 5579 + }, + { + "epoch": 77.50305676855895, + "grad_norm": 0.014171584509313107, + "learning_rate": 0.0006, + "loss": 4.225279808044434, + "step": 5580 + }, + { + "epoch": 77.5170305676856, + "grad_norm": 0.014715159311890602, + "learning_rate": 0.0006, + "loss": 4.369834899902344, + "step": 5581 + }, + { + "epoch": 77.53100436681223, + "grad_norm": 0.016554079949855804, + "learning_rate": 0.0006, + "loss": 4.2976179122924805, + "step": 5582 + }, + { + "epoch": 77.54497816593886, + "grad_norm": 0.015095747075974941, + "learning_rate": 0.0006, + "loss": 4.294407844543457, + "step": 5583 + }, + { + "epoch": 77.5589519650655, + "grad_norm": 0.012726361863315105, + "learning_rate": 0.0006, + "loss": 4.221401214599609, + "step": 5584 + }, + { + "epoch": 77.57292576419214, + "grad_norm": 0.014999309554696083, + "learning_rate": 0.0006, + "loss": 4.243929386138916, + "step": 5585 + }, + { + "epoch": 77.58689956331878, + "grad_norm": 0.016526037827134132, + "learning_rate": 0.0006, + "loss": 4.374139785766602, + "step": 5586 + }, + { + "epoch": 77.60087336244541, + "grad_norm": 0.01648339442908764, + "learning_rate": 0.0006, + "loss": 4.392425060272217, + "step": 5587 + }, + { + "epoch": 77.61484716157206, + "grad_norm": 0.015269213356077671, + "learning_rate": 0.0006, + "loss": 4.33986759185791, + "step": 5588 + }, + { + "epoch": 77.62882096069869, + "grad_norm": 0.01690507121384144, + "learning_rate": 0.0006, + "loss": 4.307476043701172, + "step": 5589 + }, + { + "epoch": 77.64279475982532, + "grad_norm": 0.016438685357570648, + "learning_rate": 0.0006, + "loss": 4.2406206130981445, + "step": 5590 + }, + { + "epoch": 77.65676855895197, + "grad_norm": 0.0157600287348032, + "learning_rate": 0.0006, + "loss": 4.314070224761963, + "step": 5591 + }, + { + "epoch": 77.6707423580786, + "grad_norm": 0.015127947553992271, + "learning_rate": 0.0006, + "loss": 4.22475528717041, + "step": 5592 + }, + { + "epoch": 77.68471615720524, + "grad_norm": 0.016632677987217903, + "learning_rate": 0.0006, + "loss": 4.26954984664917, + "step": 5593 + }, + { + "epoch": 77.69868995633188, + "grad_norm": 0.013150460086762905, + "learning_rate": 0.0006, + "loss": 4.369131088256836, + "step": 5594 + }, + { + "epoch": 77.71266375545852, + "grad_norm": 0.014513005502521992, + "learning_rate": 0.0006, + "loss": 4.3248796463012695, + "step": 5595 + }, + { + "epoch": 77.72663755458515, + "grad_norm": 0.014090991578996181, + "learning_rate": 0.0006, + "loss": 4.268381118774414, + "step": 5596 + }, + { + "epoch": 77.74061135371178, + "grad_norm": 0.01633097417652607, + "learning_rate": 0.0006, + "loss": 4.238439083099365, + "step": 5597 + }, + { + "epoch": 77.75458515283843, + "grad_norm": 0.01710933819413185, + "learning_rate": 0.0006, + "loss": 4.399697780609131, + "step": 5598 + }, + { + "epoch": 77.76855895196506, + "grad_norm": 0.019251767545938492, + "learning_rate": 0.0006, + "loss": 4.290450096130371, + "step": 5599 + }, + { + "epoch": 77.78253275109171, + "grad_norm": 0.01849171705543995, + "learning_rate": 0.0006, + "loss": 4.267734527587891, + "step": 5600 + }, + { + "epoch": 77.79650655021834, + "grad_norm": 0.014173115603625774, + "learning_rate": 0.0006, + "loss": 4.288799285888672, + "step": 5601 + }, + { + "epoch": 77.81048034934499, + "grad_norm": 0.014572718180716038, + "learning_rate": 0.0006, + "loss": 4.3790435791015625, + "step": 5602 + }, + { + "epoch": 77.82445414847162, + "grad_norm": 0.016375336796045303, + "learning_rate": 0.0006, + "loss": 4.231725692749023, + "step": 5603 + }, + { + "epoch": 77.83842794759825, + "grad_norm": 0.017326701432466507, + "learning_rate": 0.0006, + "loss": 4.211495399475098, + "step": 5604 + }, + { + "epoch": 77.8524017467249, + "grad_norm": 0.016399390995502472, + "learning_rate": 0.0006, + "loss": 4.386699199676514, + "step": 5605 + }, + { + "epoch": 77.86637554585153, + "grad_norm": 0.018269989639520645, + "learning_rate": 0.0006, + "loss": 4.254856109619141, + "step": 5606 + }, + { + "epoch": 77.88034934497817, + "grad_norm": 0.018443435430526733, + "learning_rate": 0.0006, + "loss": 4.391961097717285, + "step": 5607 + }, + { + "epoch": 77.8943231441048, + "grad_norm": 0.01784636080265045, + "learning_rate": 0.0006, + "loss": 4.273881912231445, + "step": 5608 + }, + { + "epoch": 77.90829694323143, + "grad_norm": 0.01782994158565998, + "learning_rate": 0.0006, + "loss": 4.26833963394165, + "step": 5609 + }, + { + "epoch": 77.92227074235808, + "grad_norm": 0.016260741278529167, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 5610 + }, + { + "epoch": 77.93624454148471, + "grad_norm": 0.015482903458178043, + "learning_rate": 0.0006, + "loss": 4.187826633453369, + "step": 5611 + }, + { + "epoch": 77.95021834061136, + "grad_norm": 0.015193257480859756, + "learning_rate": 0.0006, + "loss": 4.286060333251953, + "step": 5612 + }, + { + "epoch": 77.96419213973799, + "grad_norm": 0.015110115520656109, + "learning_rate": 0.0006, + "loss": 4.2828192710876465, + "step": 5613 + }, + { + "epoch": 77.97816593886463, + "grad_norm": 0.015973486006259918, + "learning_rate": 0.0006, + "loss": 4.317939758300781, + "step": 5614 + }, + { + "epoch": 77.99213973799127, + "grad_norm": 0.017363594844937325, + "learning_rate": 0.0006, + "loss": 4.287721633911133, + "step": 5615 + }, + { + "epoch": 78.0, + "grad_norm": 0.015277642756700516, + "learning_rate": 0.0006, + "loss": 4.356618881225586, + "step": 5616 + }, + { + "epoch": 78.0, + "eval_loss": 4.642678737640381, + "eval_runtime": 56.7335, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 5616 + }, + { + "epoch": 78.01397379912663, + "grad_norm": 0.016794051975011826, + "learning_rate": 0.0006, + "loss": 4.279930591583252, + "step": 5617 + }, + { + "epoch": 78.02794759825328, + "grad_norm": 0.01632673107087612, + "learning_rate": 0.0006, + "loss": 4.235943794250488, + "step": 5618 + }, + { + "epoch": 78.04192139737991, + "grad_norm": 0.01437876746058464, + "learning_rate": 0.0006, + "loss": 4.2542243003845215, + "step": 5619 + }, + { + "epoch": 78.05589519650655, + "grad_norm": 0.014993119053542614, + "learning_rate": 0.0006, + "loss": 4.262046813964844, + "step": 5620 + }, + { + "epoch": 78.06986899563319, + "grad_norm": 0.015721119940280914, + "learning_rate": 0.0006, + "loss": 4.173436164855957, + "step": 5621 + }, + { + "epoch": 78.08384279475983, + "grad_norm": 0.01334085687994957, + "learning_rate": 0.0006, + "loss": 4.322175979614258, + "step": 5622 + }, + { + "epoch": 78.09781659388646, + "grad_norm": 0.015561388805508614, + "learning_rate": 0.0006, + "loss": 4.242125988006592, + "step": 5623 + }, + { + "epoch": 78.1117903930131, + "grad_norm": 0.014418602921068668, + "learning_rate": 0.0006, + "loss": 4.266186714172363, + "step": 5624 + }, + { + "epoch": 78.12576419213974, + "grad_norm": 0.013739909045398235, + "learning_rate": 0.0006, + "loss": 4.231633186340332, + "step": 5625 + }, + { + "epoch": 78.13973799126637, + "grad_norm": 0.01402067206799984, + "learning_rate": 0.0006, + "loss": 4.188406944274902, + "step": 5626 + }, + { + "epoch": 78.15371179039302, + "grad_norm": 0.013639481738209724, + "learning_rate": 0.0006, + "loss": 4.1586594581604, + "step": 5627 + }, + { + "epoch": 78.16768558951965, + "grad_norm": 0.012858159840106964, + "learning_rate": 0.0006, + "loss": 4.360448360443115, + "step": 5628 + }, + { + "epoch": 78.18165938864628, + "grad_norm": 0.013089739717543125, + "learning_rate": 0.0006, + "loss": 4.344371795654297, + "step": 5629 + }, + { + "epoch": 78.19563318777293, + "grad_norm": 0.01382314320653677, + "learning_rate": 0.0006, + "loss": 4.20378303527832, + "step": 5630 + }, + { + "epoch": 78.20960698689956, + "grad_norm": 0.014206619001924992, + "learning_rate": 0.0006, + "loss": 4.239891529083252, + "step": 5631 + }, + { + "epoch": 78.2235807860262, + "grad_norm": 0.01472168043255806, + "learning_rate": 0.0006, + "loss": 4.282382488250732, + "step": 5632 + }, + { + "epoch": 78.23755458515284, + "grad_norm": 0.014237025752663612, + "learning_rate": 0.0006, + "loss": 4.241545677185059, + "step": 5633 + }, + { + "epoch": 78.25152838427948, + "grad_norm": 0.013231472112238407, + "learning_rate": 0.0006, + "loss": 4.184896469116211, + "step": 5634 + }, + { + "epoch": 78.26550218340611, + "grad_norm": 0.013538471423089504, + "learning_rate": 0.0006, + "loss": 4.325181007385254, + "step": 5635 + }, + { + "epoch": 78.27947598253274, + "grad_norm": 0.013541961088776588, + "learning_rate": 0.0006, + "loss": 4.368722915649414, + "step": 5636 + }, + { + "epoch": 78.29344978165939, + "grad_norm": 0.013446723110973835, + "learning_rate": 0.0006, + "loss": 4.301422595977783, + "step": 5637 + }, + { + "epoch": 78.30742358078602, + "grad_norm": 0.014753809198737144, + "learning_rate": 0.0006, + "loss": 4.259127616882324, + "step": 5638 + }, + { + "epoch": 78.32139737991267, + "grad_norm": 0.013256008736789227, + "learning_rate": 0.0006, + "loss": 4.284032821655273, + "step": 5639 + }, + { + "epoch": 78.3353711790393, + "grad_norm": 0.014272629283368587, + "learning_rate": 0.0006, + "loss": 4.234989166259766, + "step": 5640 + }, + { + "epoch": 78.34934497816595, + "grad_norm": 0.014756826683878899, + "learning_rate": 0.0006, + "loss": 4.143499374389648, + "step": 5641 + }, + { + "epoch": 78.36331877729258, + "grad_norm": 0.016549205407500267, + "learning_rate": 0.0006, + "loss": 4.311034202575684, + "step": 5642 + }, + { + "epoch": 78.37729257641921, + "grad_norm": 0.017735283821821213, + "learning_rate": 0.0006, + "loss": 4.246767520904541, + "step": 5643 + }, + { + "epoch": 78.39126637554585, + "grad_norm": 0.017370613291859627, + "learning_rate": 0.0006, + "loss": 4.2460832595825195, + "step": 5644 + }, + { + "epoch": 78.40524017467249, + "grad_norm": 0.014642846770584583, + "learning_rate": 0.0006, + "loss": 4.303811550140381, + "step": 5645 + }, + { + "epoch": 78.41921397379913, + "grad_norm": 0.013445726595818996, + "learning_rate": 0.0006, + "loss": 4.25921630859375, + "step": 5646 + }, + { + "epoch": 78.43318777292576, + "grad_norm": 0.014254065230488777, + "learning_rate": 0.0006, + "loss": 4.129446506500244, + "step": 5647 + }, + { + "epoch": 78.44716157205241, + "grad_norm": 0.01883019506931305, + "learning_rate": 0.0006, + "loss": 4.191716194152832, + "step": 5648 + }, + { + "epoch": 78.46113537117904, + "grad_norm": 0.0204437468200922, + "learning_rate": 0.0006, + "loss": 4.333085536956787, + "step": 5649 + }, + { + "epoch": 78.47510917030567, + "grad_norm": 0.01652318798005581, + "learning_rate": 0.0006, + "loss": 4.135462760925293, + "step": 5650 + }, + { + "epoch": 78.48908296943232, + "grad_norm": 0.012753551825881004, + "learning_rate": 0.0006, + "loss": 4.195051193237305, + "step": 5651 + }, + { + "epoch": 78.50305676855895, + "grad_norm": 0.01499118935316801, + "learning_rate": 0.0006, + "loss": 4.157092094421387, + "step": 5652 + }, + { + "epoch": 78.5170305676856, + "grad_norm": 0.015383945778012276, + "learning_rate": 0.0006, + "loss": 4.253818035125732, + "step": 5653 + }, + { + "epoch": 78.53100436681223, + "grad_norm": 0.015721332281827927, + "learning_rate": 0.0006, + "loss": 4.228775978088379, + "step": 5654 + }, + { + "epoch": 78.54497816593886, + "grad_norm": 0.014595463871955872, + "learning_rate": 0.0006, + "loss": 4.135197639465332, + "step": 5655 + }, + { + "epoch": 78.5589519650655, + "grad_norm": 0.015059644356369972, + "learning_rate": 0.0006, + "loss": 4.255614280700684, + "step": 5656 + }, + { + "epoch": 78.57292576419214, + "grad_norm": 0.013833344914019108, + "learning_rate": 0.0006, + "loss": 4.301250457763672, + "step": 5657 + }, + { + "epoch": 78.58689956331878, + "grad_norm": 0.013665376231074333, + "learning_rate": 0.0006, + "loss": 4.1734232902526855, + "step": 5658 + }, + { + "epoch": 78.60087336244541, + "grad_norm": 0.01539086177945137, + "learning_rate": 0.0006, + "loss": 4.221459865570068, + "step": 5659 + }, + { + "epoch": 78.61484716157206, + "grad_norm": 0.013313917443156242, + "learning_rate": 0.0006, + "loss": 4.255619049072266, + "step": 5660 + }, + { + "epoch": 78.62882096069869, + "grad_norm": 0.01447741687297821, + "learning_rate": 0.0006, + "loss": 4.298219680786133, + "step": 5661 + }, + { + "epoch": 78.64279475982532, + "grad_norm": 0.012796915136277676, + "learning_rate": 0.0006, + "loss": 4.2542405128479, + "step": 5662 + }, + { + "epoch": 78.65676855895197, + "grad_norm": 0.014058141969144344, + "learning_rate": 0.0006, + "loss": 4.284477233886719, + "step": 5663 + }, + { + "epoch": 78.6707423580786, + "grad_norm": 0.012705069966614246, + "learning_rate": 0.0006, + "loss": 4.2826666831970215, + "step": 5664 + }, + { + "epoch": 78.68471615720524, + "grad_norm": 0.013245890848338604, + "learning_rate": 0.0006, + "loss": 4.371612548828125, + "step": 5665 + }, + { + "epoch": 78.69868995633188, + "grad_norm": 0.014192677102982998, + "learning_rate": 0.0006, + "loss": 4.267344951629639, + "step": 5666 + }, + { + "epoch": 78.71266375545852, + "grad_norm": 0.015346777625381947, + "learning_rate": 0.0006, + "loss": 4.222797393798828, + "step": 5667 + }, + { + "epoch": 78.72663755458515, + "grad_norm": 0.014068621210753918, + "learning_rate": 0.0006, + "loss": 4.172964572906494, + "step": 5668 + }, + { + "epoch": 78.74061135371178, + "grad_norm": 0.014364945702254772, + "learning_rate": 0.0006, + "loss": 4.259319305419922, + "step": 5669 + }, + { + "epoch": 78.75458515283843, + "grad_norm": 0.017337387427687645, + "learning_rate": 0.0006, + "loss": 4.29533576965332, + "step": 5670 + }, + { + "epoch": 78.76855895196506, + "grad_norm": 0.017416715621948242, + "learning_rate": 0.0006, + "loss": 4.340364456176758, + "step": 5671 + }, + { + "epoch": 78.78253275109171, + "grad_norm": 0.015925846993923187, + "learning_rate": 0.0006, + "loss": 4.177454948425293, + "step": 5672 + }, + { + "epoch": 78.79650655021834, + "grad_norm": 0.013130443170666695, + "learning_rate": 0.0006, + "loss": 4.277454853057861, + "step": 5673 + }, + { + "epoch": 78.81048034934499, + "grad_norm": 0.014628687873482704, + "learning_rate": 0.0006, + "loss": 4.273757457733154, + "step": 5674 + }, + { + "epoch": 78.82445414847162, + "grad_norm": 0.01601659320294857, + "learning_rate": 0.0006, + "loss": 4.281774520874023, + "step": 5675 + }, + { + "epoch": 78.83842794759825, + "grad_norm": 0.015443485230207443, + "learning_rate": 0.0006, + "loss": 4.199937343597412, + "step": 5676 + }, + { + "epoch": 78.8524017467249, + "grad_norm": 0.013934246264398098, + "learning_rate": 0.0006, + "loss": 4.331369876861572, + "step": 5677 + }, + { + "epoch": 78.86637554585153, + "grad_norm": 0.01572083681821823, + "learning_rate": 0.0006, + "loss": 4.2828474044799805, + "step": 5678 + }, + { + "epoch": 78.88034934497817, + "grad_norm": 0.013444120064377785, + "learning_rate": 0.0006, + "loss": 4.183859825134277, + "step": 5679 + }, + { + "epoch": 78.8943231441048, + "grad_norm": 0.01591254211962223, + "learning_rate": 0.0006, + "loss": 4.295256614685059, + "step": 5680 + }, + { + "epoch": 78.90829694323143, + "grad_norm": 0.017556684091687202, + "learning_rate": 0.0006, + "loss": 4.132272720336914, + "step": 5681 + }, + { + "epoch": 78.92227074235808, + "grad_norm": 0.018184322863817215, + "learning_rate": 0.0006, + "loss": 4.251728057861328, + "step": 5682 + }, + { + "epoch": 78.93624454148471, + "grad_norm": 0.016207285225391388, + "learning_rate": 0.0006, + "loss": 4.205544471740723, + "step": 5683 + }, + { + "epoch": 78.95021834061136, + "grad_norm": 0.014478376135230064, + "learning_rate": 0.0006, + "loss": 4.182462215423584, + "step": 5684 + }, + { + "epoch": 78.96419213973799, + "grad_norm": 0.014255235902965069, + "learning_rate": 0.0006, + "loss": 4.253366947174072, + "step": 5685 + }, + { + "epoch": 78.97816593886463, + "grad_norm": 0.01524396613240242, + "learning_rate": 0.0006, + "loss": 4.221786022186279, + "step": 5686 + }, + { + "epoch": 78.99213973799127, + "grad_norm": 0.015057693235576153, + "learning_rate": 0.0006, + "loss": 4.208454608917236, + "step": 5687 + }, + { + "epoch": 79.0, + "grad_norm": 0.015478022396564484, + "learning_rate": 0.0006, + "loss": 4.336383819580078, + "step": 5688 + }, + { + "epoch": 79.0, + "eval_loss": 4.641749382019043, + "eval_runtime": 56.8167, + "eval_samples_per_second": 42.98, + "eval_steps_per_second": 1.355, + "step": 5688 + }, + { + "epoch": 79.01397379912663, + "grad_norm": 0.014143316075205803, + "learning_rate": 0.0006, + "loss": 4.260974884033203, + "step": 5689 + }, + { + "epoch": 79.02794759825328, + "grad_norm": 0.013120883144438267, + "learning_rate": 0.0006, + "loss": 4.139886856079102, + "step": 5690 + }, + { + "epoch": 79.04192139737991, + "grad_norm": 0.015019661746919155, + "learning_rate": 0.0006, + "loss": 4.267004013061523, + "step": 5691 + }, + { + "epoch": 79.05589519650655, + "grad_norm": 0.014891284517943859, + "learning_rate": 0.0006, + "loss": 4.205019950866699, + "step": 5692 + }, + { + "epoch": 79.06986899563319, + "grad_norm": 0.015527728945016861, + "learning_rate": 0.0006, + "loss": 4.1513566970825195, + "step": 5693 + }, + { + "epoch": 79.08384279475983, + "grad_norm": 0.017527181655168533, + "learning_rate": 0.0006, + "loss": 4.148828983306885, + "step": 5694 + }, + { + "epoch": 79.09781659388646, + "grad_norm": 0.01632206328213215, + "learning_rate": 0.0006, + "loss": 4.195789813995361, + "step": 5695 + }, + { + "epoch": 79.1117903930131, + "grad_norm": 0.01590503193438053, + "learning_rate": 0.0006, + "loss": 4.199155330657959, + "step": 5696 + }, + { + "epoch": 79.12576419213974, + "grad_norm": 0.016489997506141663, + "learning_rate": 0.0006, + "loss": 4.155154705047607, + "step": 5697 + }, + { + "epoch": 79.13973799126637, + "grad_norm": 0.0161368977278471, + "learning_rate": 0.0006, + "loss": 4.164384841918945, + "step": 5698 + }, + { + "epoch": 79.15371179039302, + "grad_norm": 0.018574809655547142, + "learning_rate": 0.0006, + "loss": 4.1898193359375, + "step": 5699 + }, + { + "epoch": 79.16768558951965, + "grad_norm": 0.018987983465194702, + "learning_rate": 0.0006, + "loss": 4.219719409942627, + "step": 5700 + }, + { + "epoch": 79.18165938864628, + "grad_norm": 0.01668606884777546, + "learning_rate": 0.0006, + "loss": 4.152734756469727, + "step": 5701 + }, + { + "epoch": 79.19563318777293, + "grad_norm": 0.014046299271285534, + "learning_rate": 0.0006, + "loss": 4.152964115142822, + "step": 5702 + }, + { + "epoch": 79.20960698689956, + "grad_norm": 0.01628032699227333, + "learning_rate": 0.0006, + "loss": 4.299479961395264, + "step": 5703 + }, + { + "epoch": 79.2235807860262, + "grad_norm": 0.017998795956373215, + "learning_rate": 0.0006, + "loss": 4.231203079223633, + "step": 5704 + }, + { + "epoch": 79.23755458515284, + "grad_norm": 0.014799647964537144, + "learning_rate": 0.0006, + "loss": 4.3530192375183105, + "step": 5705 + }, + { + "epoch": 79.25152838427948, + "grad_norm": 0.014640678651630878, + "learning_rate": 0.0006, + "loss": 4.287322998046875, + "step": 5706 + }, + { + "epoch": 79.26550218340611, + "grad_norm": 0.0172616895288229, + "learning_rate": 0.0006, + "loss": 4.148009777069092, + "step": 5707 + }, + { + "epoch": 79.27947598253274, + "grad_norm": 0.01811790093779564, + "learning_rate": 0.0006, + "loss": 4.269433975219727, + "step": 5708 + }, + { + "epoch": 79.29344978165939, + "grad_norm": 0.01764000579714775, + "learning_rate": 0.0006, + "loss": 4.292638778686523, + "step": 5709 + }, + { + "epoch": 79.30742358078602, + "grad_norm": 0.018030935898423195, + "learning_rate": 0.0006, + "loss": 4.267889976501465, + "step": 5710 + }, + { + "epoch": 79.32139737991267, + "grad_norm": 0.017860891297459602, + "learning_rate": 0.0006, + "loss": 4.211110591888428, + "step": 5711 + }, + { + "epoch": 79.3353711790393, + "grad_norm": 0.017295073717832565, + "learning_rate": 0.0006, + "loss": 4.224645614624023, + "step": 5712 + }, + { + "epoch": 79.34934497816595, + "grad_norm": 0.015891101211309433, + "learning_rate": 0.0006, + "loss": 4.224797248840332, + "step": 5713 + }, + { + "epoch": 79.36331877729258, + "grad_norm": 0.014680417254567146, + "learning_rate": 0.0006, + "loss": 4.182242393493652, + "step": 5714 + }, + { + "epoch": 79.37729257641921, + "grad_norm": 0.016621140763163567, + "learning_rate": 0.0006, + "loss": 4.186334609985352, + "step": 5715 + }, + { + "epoch": 79.39126637554585, + "grad_norm": 0.016207249835133553, + "learning_rate": 0.0006, + "loss": 4.277074813842773, + "step": 5716 + }, + { + "epoch": 79.40524017467249, + "grad_norm": 0.015375478193163872, + "learning_rate": 0.0006, + "loss": 4.154317855834961, + "step": 5717 + }, + { + "epoch": 79.41921397379913, + "grad_norm": 0.013453228399157524, + "learning_rate": 0.0006, + "loss": 4.265270233154297, + "step": 5718 + }, + { + "epoch": 79.43318777292576, + "grad_norm": 0.014312694780528545, + "learning_rate": 0.0006, + "loss": 4.232220649719238, + "step": 5719 + }, + { + "epoch": 79.44716157205241, + "grad_norm": 0.014759554527699947, + "learning_rate": 0.0006, + "loss": 4.272275924682617, + "step": 5720 + }, + { + "epoch": 79.46113537117904, + "grad_norm": 0.015885505825281143, + "learning_rate": 0.0006, + "loss": 4.2229719161987305, + "step": 5721 + }, + { + "epoch": 79.47510917030567, + "grad_norm": 0.016516581177711487, + "learning_rate": 0.0006, + "loss": 4.200624465942383, + "step": 5722 + }, + { + "epoch": 79.48908296943232, + "grad_norm": 0.018254555761814117, + "learning_rate": 0.0006, + "loss": 4.19648551940918, + "step": 5723 + }, + { + "epoch": 79.50305676855895, + "grad_norm": 0.01467090006917715, + "learning_rate": 0.0006, + "loss": 4.135190486907959, + "step": 5724 + }, + { + "epoch": 79.5170305676856, + "grad_norm": 0.013474797829985619, + "learning_rate": 0.0006, + "loss": 4.311361312866211, + "step": 5725 + }, + { + "epoch": 79.53100436681223, + "grad_norm": 0.01742764748632908, + "learning_rate": 0.0006, + "loss": 4.1845903396606445, + "step": 5726 + }, + { + "epoch": 79.54497816593886, + "grad_norm": 0.018786882981657982, + "learning_rate": 0.0006, + "loss": 4.163942337036133, + "step": 5727 + }, + { + "epoch": 79.5589519650655, + "grad_norm": 0.017831219360232353, + "learning_rate": 0.0006, + "loss": 4.236685276031494, + "step": 5728 + }, + { + "epoch": 79.57292576419214, + "grad_norm": 0.016334986314177513, + "learning_rate": 0.0006, + "loss": 4.255853652954102, + "step": 5729 + }, + { + "epoch": 79.58689956331878, + "grad_norm": 0.0138487434014678, + "learning_rate": 0.0006, + "loss": 4.227874279022217, + "step": 5730 + }, + { + "epoch": 79.60087336244541, + "grad_norm": 0.01531760673969984, + "learning_rate": 0.0006, + "loss": 4.221432685852051, + "step": 5731 + }, + { + "epoch": 79.61484716157206, + "grad_norm": 0.01603361964225769, + "learning_rate": 0.0006, + "loss": 4.278326511383057, + "step": 5732 + }, + { + "epoch": 79.62882096069869, + "grad_norm": 0.013149895705282688, + "learning_rate": 0.0006, + "loss": 4.2474517822265625, + "step": 5733 + }, + { + "epoch": 79.64279475982532, + "grad_norm": 0.013428786769509315, + "learning_rate": 0.0006, + "loss": 4.1701178550720215, + "step": 5734 + }, + { + "epoch": 79.65676855895197, + "grad_norm": 0.014305585995316505, + "learning_rate": 0.0006, + "loss": 4.250778675079346, + "step": 5735 + }, + { + "epoch": 79.6707423580786, + "grad_norm": 0.014544566161930561, + "learning_rate": 0.0006, + "loss": 4.243236541748047, + "step": 5736 + }, + { + "epoch": 79.68471615720524, + "grad_norm": 0.013524414040148258, + "learning_rate": 0.0006, + "loss": 4.255457878112793, + "step": 5737 + }, + { + "epoch": 79.69868995633188, + "grad_norm": 0.014750641770660877, + "learning_rate": 0.0006, + "loss": 4.202816486358643, + "step": 5738 + }, + { + "epoch": 79.71266375545852, + "grad_norm": 0.015090773813426495, + "learning_rate": 0.0006, + "loss": 4.3053178787231445, + "step": 5739 + }, + { + "epoch": 79.72663755458515, + "grad_norm": 0.01606169529259205, + "learning_rate": 0.0006, + "loss": 4.224017143249512, + "step": 5740 + }, + { + "epoch": 79.74061135371178, + "grad_norm": 0.016303520649671555, + "learning_rate": 0.0006, + "loss": 4.190760612487793, + "step": 5741 + }, + { + "epoch": 79.75458515283843, + "grad_norm": 0.01465672068297863, + "learning_rate": 0.0006, + "loss": 4.24506950378418, + "step": 5742 + }, + { + "epoch": 79.76855895196506, + "grad_norm": 0.012616301886737347, + "learning_rate": 0.0006, + "loss": 4.174039840698242, + "step": 5743 + }, + { + "epoch": 79.78253275109171, + "grad_norm": 0.01358284242451191, + "learning_rate": 0.0006, + "loss": 4.236380577087402, + "step": 5744 + }, + { + "epoch": 79.79650655021834, + "grad_norm": 0.01557177770882845, + "learning_rate": 0.0006, + "loss": 4.102962493896484, + "step": 5745 + }, + { + "epoch": 79.81048034934499, + "grad_norm": 0.015584317035973072, + "learning_rate": 0.0006, + "loss": 4.242603302001953, + "step": 5746 + }, + { + "epoch": 79.82445414847162, + "grad_norm": 0.012597334571182728, + "learning_rate": 0.0006, + "loss": 4.202791690826416, + "step": 5747 + }, + { + "epoch": 79.83842794759825, + "grad_norm": 0.013722898438572884, + "learning_rate": 0.0006, + "loss": 4.204328536987305, + "step": 5748 + }, + { + "epoch": 79.8524017467249, + "grad_norm": 0.014669225551187992, + "learning_rate": 0.0006, + "loss": 4.164331912994385, + "step": 5749 + }, + { + "epoch": 79.86637554585153, + "grad_norm": 0.015012740157544613, + "learning_rate": 0.0006, + "loss": 4.260565280914307, + "step": 5750 + }, + { + "epoch": 79.88034934497817, + "grad_norm": 0.016026340425014496, + "learning_rate": 0.0006, + "loss": 4.299723148345947, + "step": 5751 + }, + { + "epoch": 79.8943231441048, + "grad_norm": 0.01626049168407917, + "learning_rate": 0.0006, + "loss": 4.207252502441406, + "step": 5752 + }, + { + "epoch": 79.90829694323143, + "grad_norm": 0.015415354631841183, + "learning_rate": 0.0006, + "loss": 4.151934623718262, + "step": 5753 + }, + { + "epoch": 79.92227074235808, + "grad_norm": 0.014158242382109165, + "learning_rate": 0.0006, + "loss": 4.3461995124816895, + "step": 5754 + }, + { + "epoch": 79.93624454148471, + "grad_norm": 0.01585628278553486, + "learning_rate": 0.0006, + "loss": 4.334761619567871, + "step": 5755 + }, + { + "epoch": 79.95021834061136, + "grad_norm": 0.013710759580135345, + "learning_rate": 0.0006, + "loss": 4.180401802062988, + "step": 5756 + }, + { + "epoch": 79.96419213973799, + "grad_norm": 0.014159204438328743, + "learning_rate": 0.0006, + "loss": 4.322331428527832, + "step": 5757 + }, + { + "epoch": 79.97816593886463, + "grad_norm": 0.015153195708990097, + "learning_rate": 0.0006, + "loss": 4.287478923797607, + "step": 5758 + }, + { + "epoch": 79.99213973799127, + "grad_norm": 0.014100227504968643, + "learning_rate": 0.0006, + "loss": 4.231807708740234, + "step": 5759 + }, + { + "epoch": 80.0, + "grad_norm": 0.015248682349920273, + "learning_rate": 0.0006, + "loss": 4.261106014251709, + "step": 5760 + }, + { + "epoch": 80.0, + "eval_loss": 4.610264301300049, + "eval_runtime": 56.514, + "eval_samples_per_second": 43.211, + "eval_steps_per_second": 1.362, + "step": 5760 + }, + { + "epoch": 80.01397379912663, + "grad_norm": 0.014730271883308887, + "learning_rate": 0.0006, + "loss": 4.108556747436523, + "step": 5761 + }, + { + "epoch": 80.02794759825328, + "grad_norm": 0.014575716108083725, + "learning_rate": 0.0006, + "loss": 4.169658660888672, + "step": 5762 + }, + { + "epoch": 80.04192139737991, + "grad_norm": 0.014706958085298538, + "learning_rate": 0.0006, + "loss": 4.252662658691406, + "step": 5763 + }, + { + "epoch": 80.05589519650655, + "grad_norm": 0.01561224739998579, + "learning_rate": 0.0006, + "loss": 4.193892955780029, + "step": 5764 + }, + { + "epoch": 80.06986899563319, + "grad_norm": 0.014702324755489826, + "learning_rate": 0.0006, + "loss": 4.1889238357543945, + "step": 5765 + }, + { + "epoch": 80.08384279475983, + "grad_norm": 0.015565741807222366, + "learning_rate": 0.0006, + "loss": 4.243549823760986, + "step": 5766 + }, + { + "epoch": 80.09781659388646, + "grad_norm": 0.01647181063890457, + "learning_rate": 0.0006, + "loss": 4.275446891784668, + "step": 5767 + }, + { + "epoch": 80.1117903930131, + "grad_norm": 0.018177716061472893, + "learning_rate": 0.0006, + "loss": 4.12040376663208, + "step": 5768 + }, + { + "epoch": 80.12576419213974, + "grad_norm": 0.01745191030204296, + "learning_rate": 0.0006, + "loss": 4.140371322631836, + "step": 5769 + }, + { + "epoch": 80.13973799126637, + "grad_norm": 0.01568993367254734, + "learning_rate": 0.0006, + "loss": 4.258280277252197, + "step": 5770 + }, + { + "epoch": 80.15371179039302, + "grad_norm": 0.01636597141623497, + "learning_rate": 0.0006, + "loss": 4.179052352905273, + "step": 5771 + }, + { + "epoch": 80.16768558951965, + "grad_norm": 0.016839100047945976, + "learning_rate": 0.0006, + "loss": 4.127636432647705, + "step": 5772 + }, + { + "epoch": 80.18165938864628, + "grad_norm": 0.015471016988158226, + "learning_rate": 0.0006, + "loss": 4.181325912475586, + "step": 5773 + }, + { + "epoch": 80.19563318777293, + "grad_norm": 0.015836089849472046, + "learning_rate": 0.0006, + "loss": 4.142168998718262, + "step": 5774 + }, + { + "epoch": 80.20960698689956, + "grad_norm": 0.014782954007387161, + "learning_rate": 0.0006, + "loss": 4.308799743652344, + "step": 5775 + }, + { + "epoch": 80.2235807860262, + "grad_norm": 0.014040893875062466, + "learning_rate": 0.0006, + "loss": 4.216936111450195, + "step": 5776 + }, + { + "epoch": 80.23755458515284, + "grad_norm": 0.01420280709862709, + "learning_rate": 0.0006, + "loss": 4.235230445861816, + "step": 5777 + }, + { + "epoch": 80.25152838427948, + "grad_norm": 0.013341007754206657, + "learning_rate": 0.0006, + "loss": 4.369109153747559, + "step": 5778 + }, + { + "epoch": 80.26550218340611, + "grad_norm": 0.014208327978849411, + "learning_rate": 0.0006, + "loss": 4.219112396240234, + "step": 5779 + }, + { + "epoch": 80.27947598253274, + "grad_norm": 0.013825052417814732, + "learning_rate": 0.0006, + "loss": 4.1983489990234375, + "step": 5780 + }, + { + "epoch": 80.29344978165939, + "grad_norm": 0.012916515581309795, + "learning_rate": 0.0006, + "loss": 4.204582691192627, + "step": 5781 + }, + { + "epoch": 80.30742358078602, + "grad_norm": 0.013071243651211262, + "learning_rate": 0.0006, + "loss": 4.17643928527832, + "step": 5782 + }, + { + "epoch": 80.32139737991267, + "grad_norm": 0.015507174655795097, + "learning_rate": 0.0006, + "loss": 4.172976016998291, + "step": 5783 + }, + { + "epoch": 80.3353711790393, + "grad_norm": 0.014967243187129498, + "learning_rate": 0.0006, + "loss": 4.141168594360352, + "step": 5784 + }, + { + "epoch": 80.34934497816595, + "grad_norm": 0.015629924833774567, + "learning_rate": 0.0006, + "loss": 4.253526210784912, + "step": 5785 + }, + { + "epoch": 80.36331877729258, + "grad_norm": 0.01607954502105713, + "learning_rate": 0.0006, + "loss": 4.249395370483398, + "step": 5786 + }, + { + "epoch": 80.37729257641921, + "grad_norm": 0.01500706560909748, + "learning_rate": 0.0006, + "loss": 4.279012680053711, + "step": 5787 + }, + { + "epoch": 80.39126637554585, + "grad_norm": 0.014343900606036186, + "learning_rate": 0.0006, + "loss": 4.268509864807129, + "step": 5788 + }, + { + "epoch": 80.40524017467249, + "grad_norm": 0.014682197012007236, + "learning_rate": 0.0006, + "loss": 4.206334114074707, + "step": 5789 + }, + { + "epoch": 80.41921397379913, + "grad_norm": 0.014989163726568222, + "learning_rate": 0.0006, + "loss": 4.143601417541504, + "step": 5790 + }, + { + "epoch": 80.43318777292576, + "grad_norm": 0.015001608058810234, + "learning_rate": 0.0006, + "loss": 4.164111137390137, + "step": 5791 + }, + { + "epoch": 80.44716157205241, + "grad_norm": 0.015222722664475441, + "learning_rate": 0.0006, + "loss": 4.2369914054870605, + "step": 5792 + }, + { + "epoch": 80.46113537117904, + "grad_norm": 0.014126183465123177, + "learning_rate": 0.0006, + "loss": 4.2209930419921875, + "step": 5793 + }, + { + "epoch": 80.47510917030567, + "grad_norm": 0.013065959326922894, + "learning_rate": 0.0006, + "loss": 4.231184959411621, + "step": 5794 + }, + { + "epoch": 80.48908296943232, + "grad_norm": 0.013833962380886078, + "learning_rate": 0.0006, + "loss": 4.226434707641602, + "step": 5795 + }, + { + "epoch": 80.50305676855895, + "grad_norm": 0.01387125626206398, + "learning_rate": 0.0006, + "loss": 4.241328716278076, + "step": 5796 + }, + { + "epoch": 80.5170305676856, + "grad_norm": 0.014145063236355782, + "learning_rate": 0.0006, + "loss": 4.2268805503845215, + "step": 5797 + }, + { + "epoch": 80.53100436681223, + "grad_norm": 0.012990240007638931, + "learning_rate": 0.0006, + "loss": 4.1718339920043945, + "step": 5798 + }, + { + "epoch": 80.54497816593886, + "grad_norm": 0.014185411855578423, + "learning_rate": 0.0006, + "loss": 4.232397556304932, + "step": 5799 + }, + { + "epoch": 80.5589519650655, + "grad_norm": 0.014260523952543736, + "learning_rate": 0.0006, + "loss": 4.282734394073486, + "step": 5800 + }, + { + "epoch": 80.57292576419214, + "grad_norm": 0.012542120181024075, + "learning_rate": 0.0006, + "loss": 4.134551525115967, + "step": 5801 + }, + { + "epoch": 80.58689956331878, + "grad_norm": 0.013078657910227776, + "learning_rate": 0.0006, + "loss": 4.229815483093262, + "step": 5802 + }, + { + "epoch": 80.60087336244541, + "grad_norm": 0.012821056880056858, + "learning_rate": 0.0006, + "loss": 4.208278179168701, + "step": 5803 + }, + { + "epoch": 80.61484716157206, + "grad_norm": 0.012964180670678616, + "learning_rate": 0.0006, + "loss": 4.188477039337158, + "step": 5804 + }, + { + "epoch": 80.62882096069869, + "grad_norm": 0.01631746254861355, + "learning_rate": 0.0006, + "loss": 4.3214497566223145, + "step": 5805 + }, + { + "epoch": 80.64279475982532, + "grad_norm": 0.01726640947163105, + "learning_rate": 0.0006, + "loss": 4.279359817504883, + "step": 5806 + }, + { + "epoch": 80.65676855895197, + "grad_norm": 0.018778258934617043, + "learning_rate": 0.0006, + "loss": 4.294047832489014, + "step": 5807 + }, + { + "epoch": 80.6707423580786, + "grad_norm": 0.01796831749379635, + "learning_rate": 0.0006, + "loss": 4.314162254333496, + "step": 5808 + }, + { + "epoch": 80.68471615720524, + "grad_norm": 0.016495538875460625, + "learning_rate": 0.0006, + "loss": 4.273871421813965, + "step": 5809 + }, + { + "epoch": 80.69868995633188, + "grad_norm": 0.013590306043624878, + "learning_rate": 0.0006, + "loss": 4.14961051940918, + "step": 5810 + }, + { + "epoch": 80.71266375545852, + "grad_norm": 0.014670426957309246, + "learning_rate": 0.0006, + "loss": 4.211853981018066, + "step": 5811 + }, + { + "epoch": 80.72663755458515, + "grad_norm": 0.01641341857612133, + "learning_rate": 0.0006, + "loss": 4.211948871612549, + "step": 5812 + }, + { + "epoch": 80.74061135371178, + "grad_norm": 0.016365470364689827, + "learning_rate": 0.0006, + "loss": 4.200774192810059, + "step": 5813 + }, + { + "epoch": 80.75458515283843, + "grad_norm": 0.01613970287144184, + "learning_rate": 0.0006, + "loss": 4.288322448730469, + "step": 5814 + }, + { + "epoch": 80.76855895196506, + "grad_norm": 0.016742777079343796, + "learning_rate": 0.0006, + "loss": 4.23801326751709, + "step": 5815 + }, + { + "epoch": 80.78253275109171, + "grad_norm": 0.015351069159805775, + "learning_rate": 0.0006, + "loss": 4.263832092285156, + "step": 5816 + }, + { + "epoch": 80.79650655021834, + "grad_norm": 0.013435336761176586, + "learning_rate": 0.0006, + "loss": 4.160857200622559, + "step": 5817 + }, + { + "epoch": 80.81048034934499, + "grad_norm": 0.01474886853247881, + "learning_rate": 0.0006, + "loss": 4.216025352478027, + "step": 5818 + }, + { + "epoch": 80.82445414847162, + "grad_norm": 0.01586935855448246, + "learning_rate": 0.0006, + "loss": 4.256402969360352, + "step": 5819 + }, + { + "epoch": 80.83842794759825, + "grad_norm": 0.014596572145819664, + "learning_rate": 0.0006, + "loss": 4.142584800720215, + "step": 5820 + }, + { + "epoch": 80.8524017467249, + "grad_norm": 0.01501593366265297, + "learning_rate": 0.0006, + "loss": 4.250462055206299, + "step": 5821 + }, + { + "epoch": 80.86637554585153, + "grad_norm": 0.014773511327803135, + "learning_rate": 0.0006, + "loss": 4.253305912017822, + "step": 5822 + }, + { + "epoch": 80.88034934497817, + "grad_norm": 0.013246326707303524, + "learning_rate": 0.0006, + "loss": 4.170881271362305, + "step": 5823 + }, + { + "epoch": 80.8943231441048, + "grad_norm": 0.013013756833970547, + "learning_rate": 0.0006, + "loss": 4.264721870422363, + "step": 5824 + }, + { + "epoch": 80.90829694323143, + "grad_norm": 0.013155903667211533, + "learning_rate": 0.0006, + "loss": 4.274667739868164, + "step": 5825 + }, + { + "epoch": 80.92227074235808, + "grad_norm": 0.01448952779173851, + "learning_rate": 0.0006, + "loss": 4.24777889251709, + "step": 5826 + }, + { + "epoch": 80.93624454148471, + "grad_norm": 0.01415231078863144, + "learning_rate": 0.0006, + "loss": 4.22014856338501, + "step": 5827 + }, + { + "epoch": 80.95021834061136, + "grad_norm": 0.014796273782849312, + "learning_rate": 0.0006, + "loss": 4.22036600112915, + "step": 5828 + }, + { + "epoch": 80.96419213973799, + "grad_norm": 0.013005326502025127, + "learning_rate": 0.0006, + "loss": 4.276123046875, + "step": 5829 + }, + { + "epoch": 80.97816593886463, + "grad_norm": 0.012851177714765072, + "learning_rate": 0.0006, + "loss": 4.248575687408447, + "step": 5830 + }, + { + "epoch": 80.99213973799127, + "grad_norm": 0.013740002177655697, + "learning_rate": 0.0006, + "loss": 4.051140308380127, + "step": 5831 + }, + { + "epoch": 81.0, + "grad_norm": 0.015296104364097118, + "learning_rate": 0.0006, + "loss": 4.251155853271484, + "step": 5832 + }, + { + "epoch": 81.0, + "eval_loss": 4.594127655029297, + "eval_runtime": 56.0959, + "eval_samples_per_second": 43.533, + "eval_steps_per_second": 1.373, + "step": 5832 + }, + { + "epoch": 81.01397379912663, + "grad_norm": 0.015804558992385864, + "learning_rate": 0.0006, + "loss": 4.152894020080566, + "step": 5833 + }, + { + "epoch": 81.02794759825328, + "grad_norm": 0.017000854015350342, + "learning_rate": 0.0006, + "loss": 4.266067981719971, + "step": 5834 + }, + { + "epoch": 81.04192139737991, + "grad_norm": 0.016351008787751198, + "learning_rate": 0.0006, + "loss": 4.192703723907471, + "step": 5835 + }, + { + "epoch": 81.05589519650655, + "grad_norm": 0.014703895896673203, + "learning_rate": 0.0006, + "loss": 4.23554801940918, + "step": 5836 + }, + { + "epoch": 81.06986899563319, + "grad_norm": 0.01619596965610981, + "learning_rate": 0.0006, + "loss": 4.233000755310059, + "step": 5837 + }, + { + "epoch": 81.08384279475983, + "grad_norm": 0.01815493032336235, + "learning_rate": 0.0006, + "loss": 4.200490951538086, + "step": 5838 + }, + { + "epoch": 81.09781659388646, + "grad_norm": 0.018305106088519096, + "learning_rate": 0.0006, + "loss": 4.154242515563965, + "step": 5839 + }, + { + "epoch": 81.1117903930131, + "grad_norm": 0.0200856514275074, + "learning_rate": 0.0006, + "loss": 4.25820779800415, + "step": 5840 + }, + { + "epoch": 81.12576419213974, + "grad_norm": 0.02101440541446209, + "learning_rate": 0.0006, + "loss": 4.168209075927734, + "step": 5841 + }, + { + "epoch": 81.13973799126637, + "grad_norm": 0.022829681634902954, + "learning_rate": 0.0006, + "loss": 4.25560998916626, + "step": 5842 + }, + { + "epoch": 81.15371179039302, + "grad_norm": 0.0195082426071167, + "learning_rate": 0.0006, + "loss": 4.2035627365112305, + "step": 5843 + }, + { + "epoch": 81.16768558951965, + "grad_norm": 0.01716248318552971, + "learning_rate": 0.0006, + "loss": 4.11322021484375, + "step": 5844 + }, + { + "epoch": 81.18165938864628, + "grad_norm": 0.02034664712846279, + "learning_rate": 0.0006, + "loss": 4.179389953613281, + "step": 5845 + }, + { + "epoch": 81.19563318777293, + "grad_norm": 0.023578710854053497, + "learning_rate": 0.0006, + "loss": 4.265695571899414, + "step": 5846 + }, + { + "epoch": 81.20960698689956, + "grad_norm": 0.019244182854890823, + "learning_rate": 0.0006, + "loss": 4.292043209075928, + "step": 5847 + }, + { + "epoch": 81.2235807860262, + "grad_norm": 0.01577291637659073, + "learning_rate": 0.0006, + "loss": 4.250776290893555, + "step": 5848 + }, + { + "epoch": 81.23755458515284, + "grad_norm": 0.01617090217769146, + "learning_rate": 0.0006, + "loss": 4.193098068237305, + "step": 5849 + }, + { + "epoch": 81.25152838427948, + "grad_norm": 0.016090184450149536, + "learning_rate": 0.0006, + "loss": 4.112357139587402, + "step": 5850 + }, + { + "epoch": 81.26550218340611, + "grad_norm": 0.01441365759819746, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 5851 + }, + { + "epoch": 81.27947598253274, + "grad_norm": 0.014122162945568562, + "learning_rate": 0.0006, + "loss": 4.276175022125244, + "step": 5852 + }, + { + "epoch": 81.29344978165939, + "grad_norm": 0.014859487302601337, + "learning_rate": 0.0006, + "loss": 4.094699859619141, + "step": 5853 + }, + { + "epoch": 81.30742358078602, + "grad_norm": 0.013624267652630806, + "learning_rate": 0.0006, + "loss": 4.2648773193359375, + "step": 5854 + }, + { + "epoch": 81.32139737991267, + "grad_norm": 0.01417913381010294, + "learning_rate": 0.0006, + "loss": 4.228567123413086, + "step": 5855 + }, + { + "epoch": 81.3353711790393, + "grad_norm": 0.014560805633664131, + "learning_rate": 0.0006, + "loss": 4.148930549621582, + "step": 5856 + }, + { + "epoch": 81.34934497816595, + "grad_norm": 0.013662154786288738, + "learning_rate": 0.0006, + "loss": 4.192107677459717, + "step": 5857 + }, + { + "epoch": 81.36331877729258, + "grad_norm": 0.013434980995953083, + "learning_rate": 0.0006, + "loss": 4.162052154541016, + "step": 5858 + }, + { + "epoch": 81.37729257641921, + "grad_norm": 0.01300742756575346, + "learning_rate": 0.0006, + "loss": 4.262385368347168, + "step": 5859 + }, + { + "epoch": 81.39126637554585, + "grad_norm": 0.012544351629912853, + "learning_rate": 0.0006, + "loss": 4.215371131896973, + "step": 5860 + }, + { + "epoch": 81.40524017467249, + "grad_norm": 0.013663645833730698, + "learning_rate": 0.0006, + "loss": 4.247312545776367, + "step": 5861 + }, + { + "epoch": 81.41921397379913, + "grad_norm": 0.013127296231687069, + "learning_rate": 0.0006, + "loss": 4.087685585021973, + "step": 5862 + }, + { + "epoch": 81.43318777292576, + "grad_norm": 0.014081938192248344, + "learning_rate": 0.0006, + "loss": 4.272648811340332, + "step": 5863 + }, + { + "epoch": 81.44716157205241, + "grad_norm": 0.015470641665160656, + "learning_rate": 0.0006, + "loss": 4.278578758239746, + "step": 5864 + }, + { + "epoch": 81.46113537117904, + "grad_norm": 0.016524532809853554, + "learning_rate": 0.0006, + "loss": 4.118268013000488, + "step": 5865 + }, + { + "epoch": 81.47510917030567, + "grad_norm": 0.016242722049355507, + "learning_rate": 0.0006, + "loss": 4.182405471801758, + "step": 5866 + }, + { + "epoch": 81.48908296943232, + "grad_norm": 0.01557017583400011, + "learning_rate": 0.0006, + "loss": 4.229862213134766, + "step": 5867 + }, + { + "epoch": 81.50305676855895, + "grad_norm": 0.01562870666384697, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 5868 + }, + { + "epoch": 81.5170305676856, + "grad_norm": 0.018606074154376984, + "learning_rate": 0.0006, + "loss": 4.148681640625, + "step": 5869 + }, + { + "epoch": 81.53100436681223, + "grad_norm": 0.017740730196237564, + "learning_rate": 0.0006, + "loss": 4.321932792663574, + "step": 5870 + }, + { + "epoch": 81.54497816593886, + "grad_norm": 0.014517677947878838, + "learning_rate": 0.0006, + "loss": 4.22862434387207, + "step": 5871 + }, + { + "epoch": 81.5589519650655, + "grad_norm": 0.01389714702963829, + "learning_rate": 0.0006, + "loss": 4.198000907897949, + "step": 5872 + }, + { + "epoch": 81.57292576419214, + "grad_norm": 0.015348682180047035, + "learning_rate": 0.0006, + "loss": 4.293485164642334, + "step": 5873 + }, + { + "epoch": 81.58689956331878, + "grad_norm": 0.015965791419148445, + "learning_rate": 0.0006, + "loss": 4.23481559753418, + "step": 5874 + }, + { + "epoch": 81.60087336244541, + "grad_norm": 0.015662726014852524, + "learning_rate": 0.0006, + "loss": 4.158148765563965, + "step": 5875 + }, + { + "epoch": 81.61484716157206, + "grad_norm": 0.014402080327272415, + "learning_rate": 0.0006, + "loss": 4.132723808288574, + "step": 5876 + }, + { + "epoch": 81.62882096069869, + "grad_norm": 0.016846125945448875, + "learning_rate": 0.0006, + "loss": 4.1654462814331055, + "step": 5877 + }, + { + "epoch": 81.64279475982532, + "grad_norm": 0.0196097232401371, + "learning_rate": 0.0006, + "loss": 4.228450298309326, + "step": 5878 + }, + { + "epoch": 81.65676855895197, + "grad_norm": 0.016005294397473335, + "learning_rate": 0.0006, + "loss": 4.230222225189209, + "step": 5879 + }, + { + "epoch": 81.6707423580786, + "grad_norm": 0.014685059897601604, + "learning_rate": 0.0006, + "loss": 4.1367387771606445, + "step": 5880 + }, + { + "epoch": 81.68471615720524, + "grad_norm": 0.01647144928574562, + "learning_rate": 0.0006, + "loss": 4.096675395965576, + "step": 5881 + }, + { + "epoch": 81.69868995633188, + "grad_norm": 0.015978610143065453, + "learning_rate": 0.0006, + "loss": 4.232292175292969, + "step": 5882 + }, + { + "epoch": 81.71266375545852, + "grad_norm": 0.01510288193821907, + "learning_rate": 0.0006, + "loss": 4.274591445922852, + "step": 5883 + }, + { + "epoch": 81.72663755458515, + "grad_norm": 0.014942476525902748, + "learning_rate": 0.0006, + "loss": 4.161482810974121, + "step": 5884 + }, + { + "epoch": 81.74061135371178, + "grad_norm": 0.014595299027860165, + "learning_rate": 0.0006, + "loss": 4.204291343688965, + "step": 5885 + }, + { + "epoch": 81.75458515283843, + "grad_norm": 0.014952429570257664, + "learning_rate": 0.0006, + "loss": 4.242499351501465, + "step": 5886 + }, + { + "epoch": 81.76855895196506, + "grad_norm": 0.014761554077267647, + "learning_rate": 0.0006, + "loss": 4.20393180847168, + "step": 5887 + }, + { + "epoch": 81.78253275109171, + "grad_norm": 0.014749647118151188, + "learning_rate": 0.0006, + "loss": 4.227565765380859, + "step": 5888 + }, + { + "epoch": 81.79650655021834, + "grad_norm": 0.014384493231773376, + "learning_rate": 0.0006, + "loss": 4.282228946685791, + "step": 5889 + }, + { + "epoch": 81.81048034934499, + "grad_norm": 0.01260330155491829, + "learning_rate": 0.0006, + "loss": 4.2882866859436035, + "step": 5890 + }, + { + "epoch": 81.82445414847162, + "grad_norm": 0.014055311679840088, + "learning_rate": 0.0006, + "loss": 4.22351598739624, + "step": 5891 + }, + { + "epoch": 81.83842794759825, + "grad_norm": 0.013674065470695496, + "learning_rate": 0.0006, + "loss": 4.240716934204102, + "step": 5892 + }, + { + "epoch": 81.8524017467249, + "grad_norm": 0.01225196197628975, + "learning_rate": 0.0006, + "loss": 4.2958598136901855, + "step": 5893 + }, + { + "epoch": 81.86637554585153, + "grad_norm": 0.013538680970668793, + "learning_rate": 0.0006, + "loss": 4.24342679977417, + "step": 5894 + }, + { + "epoch": 81.88034934497817, + "grad_norm": 0.013151394203305244, + "learning_rate": 0.0006, + "loss": 4.081999778747559, + "step": 5895 + }, + { + "epoch": 81.8943231441048, + "grad_norm": 0.014260678552091122, + "learning_rate": 0.0006, + "loss": 4.156423568725586, + "step": 5896 + }, + { + "epoch": 81.90829694323143, + "grad_norm": 0.01329776830971241, + "learning_rate": 0.0006, + "loss": 4.1713995933532715, + "step": 5897 + }, + { + "epoch": 81.92227074235808, + "grad_norm": 0.014623276889324188, + "learning_rate": 0.0006, + "loss": 4.244891166687012, + "step": 5898 + }, + { + "epoch": 81.93624454148471, + "grad_norm": 0.015760324895381927, + "learning_rate": 0.0006, + "loss": 4.076528549194336, + "step": 5899 + }, + { + "epoch": 81.95021834061136, + "grad_norm": 0.014331766404211521, + "learning_rate": 0.0006, + "loss": 4.283037185668945, + "step": 5900 + }, + { + "epoch": 81.96419213973799, + "grad_norm": 0.012383842840790749, + "learning_rate": 0.0006, + "loss": 4.117308616638184, + "step": 5901 + }, + { + "epoch": 81.97816593886463, + "grad_norm": 0.01250660140067339, + "learning_rate": 0.0006, + "loss": 4.241250991821289, + "step": 5902 + }, + { + "epoch": 81.99213973799127, + "grad_norm": 0.012276900932192802, + "learning_rate": 0.0006, + "loss": 4.141091823577881, + "step": 5903 + }, + { + "epoch": 82.0, + "grad_norm": 0.01588737778365612, + "learning_rate": 0.0006, + "loss": 4.065532684326172, + "step": 5904 + }, + { + "epoch": 82.0, + "eval_loss": 4.612304210662842, + "eval_runtime": 57.5411, + "eval_samples_per_second": 42.439, + "eval_steps_per_second": 1.338, + "step": 5904 + }, + { + "epoch": 82.01397379912663, + "grad_norm": 0.017433544620871544, + "learning_rate": 0.0006, + "loss": 4.202497482299805, + "step": 5905 + }, + { + "epoch": 82.02794759825328, + "grad_norm": 0.017260415479540825, + "learning_rate": 0.0006, + "loss": 4.099949359893799, + "step": 5906 + }, + { + "epoch": 82.04192139737991, + "grad_norm": 0.01577908545732498, + "learning_rate": 0.0006, + "loss": 4.17806339263916, + "step": 5907 + }, + { + "epoch": 82.05589519650655, + "grad_norm": 0.016093818470835686, + "learning_rate": 0.0006, + "loss": 4.219491004943848, + "step": 5908 + }, + { + "epoch": 82.06986899563319, + "grad_norm": 0.015426302328705788, + "learning_rate": 0.0006, + "loss": 4.207982063293457, + "step": 5909 + }, + { + "epoch": 82.08384279475983, + "grad_norm": 0.01605442352592945, + "learning_rate": 0.0006, + "loss": 4.265942573547363, + "step": 5910 + }, + { + "epoch": 82.09781659388646, + "grad_norm": 0.01787225715816021, + "learning_rate": 0.0006, + "loss": 4.151535511016846, + "step": 5911 + }, + { + "epoch": 82.1117903930131, + "grad_norm": 0.01750164106488228, + "learning_rate": 0.0006, + "loss": 4.133655071258545, + "step": 5912 + }, + { + "epoch": 82.12576419213974, + "grad_norm": 0.01602957211434841, + "learning_rate": 0.0006, + "loss": 4.176972389221191, + "step": 5913 + }, + { + "epoch": 82.13973799126637, + "grad_norm": 0.01553812064230442, + "learning_rate": 0.0006, + "loss": 4.125692367553711, + "step": 5914 + }, + { + "epoch": 82.15371179039302, + "grad_norm": 0.013978508301079273, + "learning_rate": 0.0006, + "loss": 4.207271575927734, + "step": 5915 + }, + { + "epoch": 82.16768558951965, + "grad_norm": 0.013561118394136429, + "learning_rate": 0.0006, + "loss": 4.238930702209473, + "step": 5916 + }, + { + "epoch": 82.18165938864628, + "grad_norm": 0.013285507448017597, + "learning_rate": 0.0006, + "loss": 4.129464626312256, + "step": 5917 + }, + { + "epoch": 82.19563318777293, + "grad_norm": 0.013903839513659477, + "learning_rate": 0.0006, + "loss": 4.214109420776367, + "step": 5918 + }, + { + "epoch": 82.20960698689956, + "grad_norm": 0.014426837675273418, + "learning_rate": 0.0006, + "loss": 4.143196105957031, + "step": 5919 + }, + { + "epoch": 82.2235807860262, + "grad_norm": 0.015347165986895561, + "learning_rate": 0.0006, + "loss": 4.235439300537109, + "step": 5920 + }, + { + "epoch": 82.23755458515284, + "grad_norm": 0.014266039244830608, + "learning_rate": 0.0006, + "loss": 3.991617202758789, + "step": 5921 + }, + { + "epoch": 82.25152838427948, + "grad_norm": 0.013101593591272831, + "learning_rate": 0.0006, + "loss": 4.266017913818359, + "step": 5922 + }, + { + "epoch": 82.26550218340611, + "grad_norm": 0.013741881586611271, + "learning_rate": 0.0006, + "loss": 4.224879741668701, + "step": 5923 + }, + { + "epoch": 82.27947598253274, + "grad_norm": 0.014206945896148682, + "learning_rate": 0.0006, + "loss": 4.102272987365723, + "step": 5924 + }, + { + "epoch": 82.29344978165939, + "grad_norm": 0.013798723928630352, + "learning_rate": 0.0006, + "loss": 4.30418586730957, + "step": 5925 + }, + { + "epoch": 82.30742358078602, + "grad_norm": 0.015470122918486595, + "learning_rate": 0.0006, + "loss": 4.2513322830200195, + "step": 5926 + }, + { + "epoch": 82.32139737991267, + "grad_norm": 0.01655442826449871, + "learning_rate": 0.0006, + "loss": 4.223002910614014, + "step": 5927 + }, + { + "epoch": 82.3353711790393, + "grad_norm": 0.017170004546642303, + "learning_rate": 0.0006, + "loss": 4.253904819488525, + "step": 5928 + }, + { + "epoch": 82.34934497816595, + "grad_norm": 0.017595812678337097, + "learning_rate": 0.0006, + "loss": 4.184927940368652, + "step": 5929 + }, + { + "epoch": 82.36331877729258, + "grad_norm": 0.015073256567120552, + "learning_rate": 0.0006, + "loss": 4.202239036560059, + "step": 5930 + }, + { + "epoch": 82.37729257641921, + "grad_norm": 0.014310941100120544, + "learning_rate": 0.0006, + "loss": 4.124934673309326, + "step": 5931 + }, + { + "epoch": 82.39126637554585, + "grad_norm": 0.013671821914613247, + "learning_rate": 0.0006, + "loss": 4.138919353485107, + "step": 5932 + }, + { + "epoch": 82.40524017467249, + "grad_norm": 0.014103314839303493, + "learning_rate": 0.0006, + "loss": 4.211289405822754, + "step": 5933 + }, + { + "epoch": 82.41921397379913, + "grad_norm": 0.014272668398916721, + "learning_rate": 0.0006, + "loss": 4.06290340423584, + "step": 5934 + }, + { + "epoch": 82.43318777292576, + "grad_norm": 0.01387016475200653, + "learning_rate": 0.0006, + "loss": 4.101334571838379, + "step": 5935 + }, + { + "epoch": 82.44716157205241, + "grad_norm": 0.013318683952093124, + "learning_rate": 0.0006, + "loss": 4.206409931182861, + "step": 5936 + }, + { + "epoch": 82.46113537117904, + "grad_norm": 0.013781268149614334, + "learning_rate": 0.0006, + "loss": 4.189474105834961, + "step": 5937 + }, + { + "epoch": 82.47510917030567, + "grad_norm": 0.013195851817727089, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 5938 + }, + { + "epoch": 82.48908296943232, + "grad_norm": 0.014358163811266422, + "learning_rate": 0.0006, + "loss": 4.203262805938721, + "step": 5939 + }, + { + "epoch": 82.50305676855895, + "grad_norm": 0.01403636671602726, + "learning_rate": 0.0006, + "loss": 4.079258441925049, + "step": 5940 + }, + { + "epoch": 82.5170305676856, + "grad_norm": 0.013272535055875778, + "learning_rate": 0.0006, + "loss": 4.19197940826416, + "step": 5941 + }, + { + "epoch": 82.53100436681223, + "grad_norm": 0.013787264935672283, + "learning_rate": 0.0006, + "loss": 4.12995719909668, + "step": 5942 + }, + { + "epoch": 82.54497816593886, + "grad_norm": 0.015020057559013367, + "learning_rate": 0.0006, + "loss": 4.148890495300293, + "step": 5943 + }, + { + "epoch": 82.5589519650655, + "grad_norm": 0.014199157245457172, + "learning_rate": 0.0006, + "loss": 4.201516151428223, + "step": 5944 + }, + { + "epoch": 82.57292576419214, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.240894317626953, + "step": 5945 + }, + { + "epoch": 82.58689956331878, + "grad_norm": 0.014767560176551342, + "learning_rate": 0.0006, + "loss": 4.173320770263672, + "step": 5946 + }, + { + "epoch": 82.60087336244541, + "grad_norm": 0.01612078957259655, + "learning_rate": 0.0006, + "loss": 4.266448020935059, + "step": 5947 + }, + { + "epoch": 82.61484716157206, + "grad_norm": 0.016706543043255806, + "learning_rate": 0.0006, + "loss": 4.229398250579834, + "step": 5948 + }, + { + "epoch": 82.62882096069869, + "grad_norm": 0.015630796551704407, + "learning_rate": 0.0006, + "loss": 4.181282997131348, + "step": 5949 + }, + { + "epoch": 82.64279475982532, + "grad_norm": 0.013103844597935677, + "learning_rate": 0.0006, + "loss": 4.309617042541504, + "step": 5950 + }, + { + "epoch": 82.65676855895197, + "grad_norm": 0.016136541962623596, + "learning_rate": 0.0006, + "loss": 4.224159240722656, + "step": 5951 + }, + { + "epoch": 82.6707423580786, + "grad_norm": 0.016160814091563225, + "learning_rate": 0.0006, + "loss": 4.267970085144043, + "step": 5952 + }, + { + "epoch": 82.68471615720524, + "grad_norm": 0.01672680303454399, + "learning_rate": 0.0006, + "loss": 4.230506896972656, + "step": 5953 + }, + { + "epoch": 82.69868995633188, + "grad_norm": 0.0159586351364851, + "learning_rate": 0.0006, + "loss": 4.159542083740234, + "step": 5954 + }, + { + "epoch": 82.71266375545852, + "grad_norm": 0.01461219135671854, + "learning_rate": 0.0006, + "loss": 4.27863073348999, + "step": 5955 + }, + { + "epoch": 82.72663755458515, + "grad_norm": 0.015088613145053387, + "learning_rate": 0.0006, + "loss": 4.272418975830078, + "step": 5956 + }, + { + "epoch": 82.74061135371178, + "grad_norm": 0.01594499684870243, + "learning_rate": 0.0006, + "loss": 4.271326065063477, + "step": 5957 + }, + { + "epoch": 82.75458515283843, + "grad_norm": 0.016648447141051292, + "learning_rate": 0.0006, + "loss": 4.141478538513184, + "step": 5958 + }, + { + "epoch": 82.76855895196506, + "grad_norm": 0.01540316641330719, + "learning_rate": 0.0006, + "loss": 4.111814498901367, + "step": 5959 + }, + { + "epoch": 82.78253275109171, + "grad_norm": 0.01634364202618599, + "learning_rate": 0.0006, + "loss": 4.3113112449646, + "step": 5960 + }, + { + "epoch": 82.79650655021834, + "grad_norm": 0.014768977649509907, + "learning_rate": 0.0006, + "loss": 4.184217929840088, + "step": 5961 + }, + { + "epoch": 82.81048034934499, + "grad_norm": 0.013453755527734756, + "learning_rate": 0.0006, + "loss": 4.247729301452637, + "step": 5962 + }, + { + "epoch": 82.82445414847162, + "grad_norm": 0.014731907285749912, + "learning_rate": 0.0006, + "loss": 4.352839469909668, + "step": 5963 + }, + { + "epoch": 82.83842794759825, + "grad_norm": 0.016047345474362373, + "learning_rate": 0.0006, + "loss": 4.257283687591553, + "step": 5964 + }, + { + "epoch": 82.8524017467249, + "grad_norm": 0.017141170799732208, + "learning_rate": 0.0006, + "loss": 4.22850227355957, + "step": 5965 + }, + { + "epoch": 82.86637554585153, + "grad_norm": 0.01370396837592125, + "learning_rate": 0.0006, + "loss": 4.1169257164001465, + "step": 5966 + }, + { + "epoch": 82.88034934497817, + "grad_norm": 0.012060081586241722, + "learning_rate": 0.0006, + "loss": 4.195242404937744, + "step": 5967 + }, + { + "epoch": 82.8943231441048, + "grad_norm": 0.015528378076851368, + "learning_rate": 0.0006, + "loss": 4.294939994812012, + "step": 5968 + }, + { + "epoch": 82.90829694323143, + "grad_norm": 0.018151041120290756, + "learning_rate": 0.0006, + "loss": 4.137192726135254, + "step": 5969 + }, + { + "epoch": 82.92227074235808, + "grad_norm": 0.017895735800266266, + "learning_rate": 0.0006, + "loss": 4.174036026000977, + "step": 5970 + }, + { + "epoch": 82.93624454148471, + "grad_norm": 0.015610386617481709, + "learning_rate": 0.0006, + "loss": 4.230136871337891, + "step": 5971 + }, + { + "epoch": 82.95021834061136, + "grad_norm": 0.01339928898960352, + "learning_rate": 0.0006, + "loss": 4.199074745178223, + "step": 5972 + }, + { + "epoch": 82.96419213973799, + "grad_norm": 0.015454214997589588, + "learning_rate": 0.0006, + "loss": 4.216109752655029, + "step": 5973 + }, + { + "epoch": 82.97816593886463, + "grad_norm": 0.014978132210671902, + "learning_rate": 0.0006, + "loss": 4.131511688232422, + "step": 5974 + }, + { + "epoch": 82.99213973799127, + "grad_norm": 0.015611122362315655, + "learning_rate": 0.0006, + "loss": 4.246518135070801, + "step": 5975 + }, + { + "epoch": 83.0, + "grad_norm": 0.016682764515280724, + "learning_rate": 0.0006, + "loss": 4.110680103302002, + "step": 5976 + }, + { + "epoch": 83.0, + "eval_loss": 4.641707897186279, + "eval_runtime": 57.2799, + "eval_samples_per_second": 42.633, + "eval_steps_per_second": 1.344, + "step": 5976 + }, + { + "epoch": 83.01397379912663, + "grad_norm": 0.015007982030510902, + "learning_rate": 0.0006, + "loss": 4.100912094116211, + "step": 5977 + }, + { + "epoch": 83.02794759825328, + "grad_norm": 0.017609436064958572, + "learning_rate": 0.0006, + "loss": 4.233399868011475, + "step": 5978 + }, + { + "epoch": 83.04192139737991, + "grad_norm": 0.019817272201180458, + "learning_rate": 0.0006, + "loss": 4.191173076629639, + "step": 5979 + }, + { + "epoch": 83.05589519650655, + "grad_norm": 0.017752140760421753, + "learning_rate": 0.0006, + "loss": 4.044437885284424, + "step": 5980 + }, + { + "epoch": 83.06986899563319, + "grad_norm": 0.015482760034501553, + "learning_rate": 0.0006, + "loss": 4.319931983947754, + "step": 5981 + }, + { + "epoch": 83.08384279475983, + "grad_norm": 0.015153988264501095, + "learning_rate": 0.0006, + "loss": 4.095277309417725, + "step": 5982 + }, + { + "epoch": 83.09781659388646, + "grad_norm": 0.017025282606482506, + "learning_rate": 0.0006, + "loss": 4.235110759735107, + "step": 5983 + }, + { + "epoch": 83.1117903930131, + "grad_norm": 0.01566586270928383, + "learning_rate": 0.0006, + "loss": 4.1934685707092285, + "step": 5984 + }, + { + "epoch": 83.12576419213974, + "grad_norm": 0.016419263556599617, + "learning_rate": 0.0006, + "loss": 4.234419822692871, + "step": 5985 + }, + { + "epoch": 83.13973799126637, + "grad_norm": 0.017132675275206566, + "learning_rate": 0.0006, + "loss": 4.221447944641113, + "step": 5986 + }, + { + "epoch": 83.15371179039302, + "grad_norm": 0.016138629987835884, + "learning_rate": 0.0006, + "loss": 4.199830055236816, + "step": 5987 + }, + { + "epoch": 83.16768558951965, + "grad_norm": 0.016210174188017845, + "learning_rate": 0.0006, + "loss": 4.081717491149902, + "step": 5988 + }, + { + "epoch": 83.18165938864628, + "grad_norm": 0.016055308282375336, + "learning_rate": 0.0006, + "loss": 4.247655391693115, + "step": 5989 + }, + { + "epoch": 83.19563318777293, + "grad_norm": 0.016132952645421028, + "learning_rate": 0.0006, + "loss": 4.262535572052002, + "step": 5990 + }, + { + "epoch": 83.20960698689956, + "grad_norm": 0.01779935695230961, + "learning_rate": 0.0006, + "loss": 4.153120040893555, + "step": 5991 + }, + { + "epoch": 83.2235807860262, + "grad_norm": 0.018360646441578865, + "learning_rate": 0.0006, + "loss": 4.232644081115723, + "step": 5992 + }, + { + "epoch": 83.23755458515284, + "grad_norm": 0.019629675894975662, + "learning_rate": 0.0006, + "loss": 3.9758734703063965, + "step": 5993 + }, + { + "epoch": 83.25152838427948, + "grad_norm": 0.017819996923208237, + "learning_rate": 0.0006, + "loss": 4.196932315826416, + "step": 5994 + }, + { + "epoch": 83.26550218340611, + "grad_norm": 0.01557681243866682, + "learning_rate": 0.0006, + "loss": 4.122654914855957, + "step": 5995 + }, + { + "epoch": 83.27947598253274, + "grad_norm": 0.017858853563666344, + "learning_rate": 0.0006, + "loss": 4.185330867767334, + "step": 5996 + }, + { + "epoch": 83.29344978165939, + "grad_norm": 0.018065953627228737, + "learning_rate": 0.0006, + "loss": 4.218471050262451, + "step": 5997 + }, + { + "epoch": 83.30742358078602, + "grad_norm": 0.017247939482331276, + "learning_rate": 0.0006, + "loss": 4.127076148986816, + "step": 5998 + }, + { + "epoch": 83.32139737991267, + "grad_norm": 0.017047952860593796, + "learning_rate": 0.0006, + "loss": 4.131356716156006, + "step": 5999 + }, + { + "epoch": 83.3353711790393, + "grad_norm": 0.015905629843473434, + "learning_rate": 0.0006, + "loss": 4.2239789962768555, + "step": 6000 + }, + { + "epoch": 83.34934497816595, + "grad_norm": 0.016522347927093506, + "learning_rate": 0.0006, + "loss": 4.054897785186768, + "step": 6001 + }, + { + "epoch": 83.36331877729258, + "grad_norm": 0.01622571051120758, + "learning_rate": 0.0006, + "loss": 4.284210681915283, + "step": 6002 + }, + { + "epoch": 83.37729257641921, + "grad_norm": 0.01640467159450054, + "learning_rate": 0.0006, + "loss": 4.268840789794922, + "step": 6003 + }, + { + "epoch": 83.39126637554585, + "grad_norm": 0.015698859468102455, + "learning_rate": 0.0006, + "loss": 4.179793357849121, + "step": 6004 + }, + { + "epoch": 83.40524017467249, + "grad_norm": 0.016694528982043266, + "learning_rate": 0.0006, + "loss": 4.132297515869141, + "step": 6005 + }, + { + "epoch": 83.41921397379913, + "grad_norm": 0.015345900319516659, + "learning_rate": 0.0006, + "loss": 4.090163230895996, + "step": 6006 + }, + { + "epoch": 83.43318777292576, + "grad_norm": 0.014722758904099464, + "learning_rate": 0.0006, + "loss": 4.179166793823242, + "step": 6007 + }, + { + "epoch": 83.44716157205241, + "grad_norm": 0.014884887263178825, + "learning_rate": 0.0006, + "loss": 4.137659072875977, + "step": 6008 + }, + { + "epoch": 83.46113537117904, + "grad_norm": 0.014041902497410774, + "learning_rate": 0.0006, + "loss": 4.1202073097229, + "step": 6009 + }, + { + "epoch": 83.47510917030567, + "grad_norm": 0.012799333781003952, + "learning_rate": 0.0006, + "loss": 4.092607498168945, + "step": 6010 + }, + { + "epoch": 83.48908296943232, + "grad_norm": 0.013272707350552082, + "learning_rate": 0.0006, + "loss": 4.132715225219727, + "step": 6011 + }, + { + "epoch": 83.50305676855895, + "grad_norm": 0.012934371829032898, + "learning_rate": 0.0006, + "loss": 4.230278968811035, + "step": 6012 + }, + { + "epoch": 83.5170305676856, + "grad_norm": 0.0120006762444973, + "learning_rate": 0.0006, + "loss": 4.140068531036377, + "step": 6013 + }, + { + "epoch": 83.53100436681223, + "grad_norm": 0.013616700656712055, + "learning_rate": 0.0006, + "loss": 4.20107364654541, + "step": 6014 + }, + { + "epoch": 83.54497816593886, + "grad_norm": 0.013957851566374302, + "learning_rate": 0.0006, + "loss": 4.124945163726807, + "step": 6015 + }, + { + "epoch": 83.5589519650655, + "grad_norm": 0.012762055732309818, + "learning_rate": 0.0006, + "loss": 4.022052764892578, + "step": 6016 + }, + { + "epoch": 83.57292576419214, + "grad_norm": 0.013080576434731483, + "learning_rate": 0.0006, + "loss": 4.103252410888672, + "step": 6017 + }, + { + "epoch": 83.58689956331878, + "grad_norm": 0.013957500457763672, + "learning_rate": 0.0006, + "loss": 4.235089302062988, + "step": 6018 + }, + { + "epoch": 83.60087336244541, + "grad_norm": 0.014707164838910103, + "learning_rate": 0.0006, + "loss": 4.191570281982422, + "step": 6019 + }, + { + "epoch": 83.61484716157206, + "grad_norm": 0.012807869352400303, + "learning_rate": 0.0006, + "loss": 4.294543266296387, + "step": 6020 + }, + { + "epoch": 83.62882096069869, + "grad_norm": 0.012027994729578495, + "learning_rate": 0.0006, + "loss": 4.13849401473999, + "step": 6021 + }, + { + "epoch": 83.64279475982532, + "grad_norm": 0.012147724628448486, + "learning_rate": 0.0006, + "loss": 4.127481937408447, + "step": 6022 + }, + { + "epoch": 83.65676855895197, + "grad_norm": 0.013241392560303211, + "learning_rate": 0.0006, + "loss": 4.218972682952881, + "step": 6023 + }, + { + "epoch": 83.6707423580786, + "grad_norm": 0.014071648940443993, + "learning_rate": 0.0006, + "loss": 4.096724033355713, + "step": 6024 + }, + { + "epoch": 83.68471615720524, + "grad_norm": 0.015212797559797764, + "learning_rate": 0.0006, + "loss": 4.27097225189209, + "step": 6025 + }, + { + "epoch": 83.69868995633188, + "grad_norm": 0.017647385597229004, + "learning_rate": 0.0006, + "loss": 4.209835529327393, + "step": 6026 + }, + { + "epoch": 83.71266375545852, + "grad_norm": 0.01709647662937641, + "learning_rate": 0.0006, + "loss": 4.042106628417969, + "step": 6027 + }, + { + "epoch": 83.72663755458515, + "grad_norm": 0.014530927874147892, + "learning_rate": 0.0006, + "loss": 4.181731700897217, + "step": 6028 + }, + { + "epoch": 83.74061135371178, + "grad_norm": 0.015356711111962795, + "learning_rate": 0.0006, + "loss": 4.314183235168457, + "step": 6029 + }, + { + "epoch": 83.75458515283843, + "grad_norm": 0.015500029549002647, + "learning_rate": 0.0006, + "loss": 4.06776762008667, + "step": 6030 + }, + { + "epoch": 83.76855895196506, + "grad_norm": 0.015351332724094391, + "learning_rate": 0.0006, + "loss": 4.185141563415527, + "step": 6031 + }, + { + "epoch": 83.78253275109171, + "grad_norm": 0.016678154468536377, + "learning_rate": 0.0006, + "loss": 4.179529666900635, + "step": 6032 + }, + { + "epoch": 83.79650655021834, + "grad_norm": 0.0156265702098608, + "learning_rate": 0.0006, + "loss": 4.238584041595459, + "step": 6033 + }, + { + "epoch": 83.81048034934499, + "grad_norm": 0.016418596729636192, + "learning_rate": 0.0006, + "loss": 4.1593523025512695, + "step": 6034 + }, + { + "epoch": 83.82445414847162, + "grad_norm": 0.016113586723804474, + "learning_rate": 0.0006, + "loss": 4.157470226287842, + "step": 6035 + }, + { + "epoch": 83.83842794759825, + "grad_norm": 0.014434042386710644, + "learning_rate": 0.0006, + "loss": 4.196226119995117, + "step": 6036 + }, + { + "epoch": 83.8524017467249, + "grad_norm": 0.012647042982280254, + "learning_rate": 0.0006, + "loss": 4.130589485168457, + "step": 6037 + }, + { + "epoch": 83.86637554585153, + "grad_norm": 0.014412877149879932, + "learning_rate": 0.0006, + "loss": 4.154845237731934, + "step": 6038 + }, + { + "epoch": 83.88034934497817, + "grad_norm": 0.014222027733922005, + "learning_rate": 0.0006, + "loss": 4.189724922180176, + "step": 6039 + }, + { + "epoch": 83.8943231441048, + "grad_norm": 0.013084255158901215, + "learning_rate": 0.0006, + "loss": 4.303338527679443, + "step": 6040 + }, + { + "epoch": 83.90829694323143, + "grad_norm": 0.012659232132136822, + "learning_rate": 0.0006, + "loss": 4.08192253112793, + "step": 6041 + }, + { + "epoch": 83.92227074235808, + "grad_norm": 0.01246276218444109, + "learning_rate": 0.0006, + "loss": 4.283324718475342, + "step": 6042 + }, + { + "epoch": 83.93624454148471, + "grad_norm": 0.01340118795633316, + "learning_rate": 0.0006, + "loss": 4.117908477783203, + "step": 6043 + }, + { + "epoch": 83.95021834061136, + "grad_norm": 0.014496722258627415, + "learning_rate": 0.0006, + "loss": 4.110318660736084, + "step": 6044 + }, + { + "epoch": 83.96419213973799, + "grad_norm": 0.013348298147320747, + "learning_rate": 0.0006, + "loss": 4.185139179229736, + "step": 6045 + }, + { + "epoch": 83.97816593886463, + "grad_norm": 0.013294277712702751, + "learning_rate": 0.0006, + "loss": 4.271645545959473, + "step": 6046 + }, + { + "epoch": 83.99213973799127, + "grad_norm": 0.014431594870984554, + "learning_rate": 0.0006, + "loss": 4.161383152008057, + "step": 6047 + }, + { + "epoch": 84.0, + "grad_norm": 0.016268927603960037, + "learning_rate": 0.0006, + "loss": 4.178324222564697, + "step": 6048 + }, + { + "epoch": 84.0, + "eval_loss": 4.6307549476623535, + "eval_runtime": 56.8521, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 6048 + }, + { + "epoch": 84.01397379912663, + "grad_norm": 0.014864908531308174, + "learning_rate": 0.0006, + "loss": 4.121300220489502, + "step": 6049 + }, + { + "epoch": 84.02794759825328, + "grad_norm": 0.014863367192447186, + "learning_rate": 0.0006, + "loss": 4.1650543212890625, + "step": 6050 + }, + { + "epoch": 84.04192139737991, + "grad_norm": 0.014379646629095078, + "learning_rate": 0.0006, + "loss": 4.0201239585876465, + "step": 6051 + }, + { + "epoch": 84.05589519650655, + "grad_norm": 0.014871201477944851, + "learning_rate": 0.0006, + "loss": 4.0820698738098145, + "step": 6052 + }, + { + "epoch": 84.06986899563319, + "grad_norm": 0.013085179962217808, + "learning_rate": 0.0006, + "loss": 4.223984718322754, + "step": 6053 + }, + { + "epoch": 84.08384279475983, + "grad_norm": 0.012485949322581291, + "learning_rate": 0.0006, + "loss": 4.182206153869629, + "step": 6054 + }, + { + "epoch": 84.09781659388646, + "grad_norm": 0.013653291389346123, + "learning_rate": 0.0006, + "loss": 4.142557144165039, + "step": 6055 + }, + { + "epoch": 84.1117903930131, + "grad_norm": 0.013930532149970531, + "learning_rate": 0.0006, + "loss": 4.13972282409668, + "step": 6056 + }, + { + "epoch": 84.12576419213974, + "grad_norm": 0.01297379657626152, + "learning_rate": 0.0006, + "loss": 4.144664764404297, + "step": 6057 + }, + { + "epoch": 84.13973799126637, + "grad_norm": 0.013210492208600044, + "learning_rate": 0.0006, + "loss": 4.253764629364014, + "step": 6058 + }, + { + "epoch": 84.15371179039302, + "grad_norm": 0.014650176279246807, + "learning_rate": 0.0006, + "loss": 4.246866226196289, + "step": 6059 + }, + { + "epoch": 84.16768558951965, + "grad_norm": 0.01482292078435421, + "learning_rate": 0.0006, + "loss": 4.140187740325928, + "step": 6060 + }, + { + "epoch": 84.18165938864628, + "grad_norm": 0.01592552848160267, + "learning_rate": 0.0006, + "loss": 4.170705318450928, + "step": 6061 + }, + { + "epoch": 84.19563318777293, + "grad_norm": 0.017152156680822372, + "learning_rate": 0.0006, + "loss": 4.126092433929443, + "step": 6062 + }, + { + "epoch": 84.20960698689956, + "grad_norm": 0.015880905091762543, + "learning_rate": 0.0006, + "loss": 4.161938190460205, + "step": 6063 + }, + { + "epoch": 84.2235807860262, + "grad_norm": 0.015451471321284771, + "learning_rate": 0.0006, + "loss": 4.263461112976074, + "step": 6064 + }, + { + "epoch": 84.23755458515284, + "grad_norm": 0.014365996234118938, + "learning_rate": 0.0006, + "loss": 4.14488410949707, + "step": 6065 + }, + { + "epoch": 84.25152838427948, + "grad_norm": 0.016364749521017075, + "learning_rate": 0.0006, + "loss": 4.112489700317383, + "step": 6066 + }, + { + "epoch": 84.26550218340611, + "grad_norm": 0.01791754551231861, + "learning_rate": 0.0006, + "loss": 4.133513450622559, + "step": 6067 + }, + { + "epoch": 84.27947598253274, + "grad_norm": 0.016609208658337593, + "learning_rate": 0.0006, + "loss": 4.201829433441162, + "step": 6068 + }, + { + "epoch": 84.29344978165939, + "grad_norm": 0.017587820068001747, + "learning_rate": 0.0006, + "loss": 4.203151702880859, + "step": 6069 + }, + { + "epoch": 84.30742358078602, + "grad_norm": 0.017047742381691933, + "learning_rate": 0.0006, + "loss": 4.077008247375488, + "step": 6070 + }, + { + "epoch": 84.32139737991267, + "grad_norm": 0.01524646207690239, + "learning_rate": 0.0006, + "loss": 4.198038101196289, + "step": 6071 + }, + { + "epoch": 84.3353711790393, + "grad_norm": 0.014196421951055527, + "learning_rate": 0.0006, + "loss": 4.164245128631592, + "step": 6072 + }, + { + "epoch": 84.34934497816595, + "grad_norm": 0.016291776672005653, + "learning_rate": 0.0006, + "loss": 4.150152206420898, + "step": 6073 + }, + { + "epoch": 84.36331877729258, + "grad_norm": 0.015627920627593994, + "learning_rate": 0.0006, + "loss": 4.110006332397461, + "step": 6074 + }, + { + "epoch": 84.37729257641921, + "grad_norm": 0.013251487165689468, + "learning_rate": 0.0006, + "loss": 4.101463317871094, + "step": 6075 + }, + { + "epoch": 84.39126637554585, + "grad_norm": 0.013601605780422688, + "learning_rate": 0.0006, + "loss": 4.191105842590332, + "step": 6076 + }, + { + "epoch": 84.40524017467249, + "grad_norm": 0.014878741465508938, + "learning_rate": 0.0006, + "loss": 4.153046131134033, + "step": 6077 + }, + { + "epoch": 84.41921397379913, + "grad_norm": 0.014783897437155247, + "learning_rate": 0.0006, + "loss": 4.362000465393066, + "step": 6078 + }, + { + "epoch": 84.43318777292576, + "grad_norm": 0.014221332035958767, + "learning_rate": 0.0006, + "loss": 4.144364833831787, + "step": 6079 + }, + { + "epoch": 84.44716157205241, + "grad_norm": 0.014911034144461155, + "learning_rate": 0.0006, + "loss": 4.088857650756836, + "step": 6080 + }, + { + "epoch": 84.46113537117904, + "grad_norm": 0.014924770221114159, + "learning_rate": 0.0006, + "loss": 4.1677446365356445, + "step": 6081 + }, + { + "epoch": 84.47510917030567, + "grad_norm": 0.014394409023225307, + "learning_rate": 0.0006, + "loss": 4.087103843688965, + "step": 6082 + }, + { + "epoch": 84.48908296943232, + "grad_norm": 0.013926676474511623, + "learning_rate": 0.0006, + "loss": 3.9972851276397705, + "step": 6083 + }, + { + "epoch": 84.50305676855895, + "grad_norm": 0.013069954700767994, + "learning_rate": 0.0006, + "loss": 3.998237133026123, + "step": 6084 + }, + { + "epoch": 84.5170305676856, + "grad_norm": 0.012748011387884617, + "learning_rate": 0.0006, + "loss": 4.195858955383301, + "step": 6085 + }, + { + "epoch": 84.53100436681223, + "grad_norm": 0.012278077192604542, + "learning_rate": 0.0006, + "loss": 4.102165222167969, + "step": 6086 + }, + { + "epoch": 84.54497816593886, + "grad_norm": 0.014259163290262222, + "learning_rate": 0.0006, + "loss": 4.092750549316406, + "step": 6087 + }, + { + "epoch": 84.5589519650655, + "grad_norm": 0.014553683809936047, + "learning_rate": 0.0006, + "loss": 4.1412763595581055, + "step": 6088 + }, + { + "epoch": 84.57292576419214, + "grad_norm": 0.013605243526399136, + "learning_rate": 0.0006, + "loss": 4.192173480987549, + "step": 6089 + }, + { + "epoch": 84.58689956331878, + "grad_norm": 0.014259358868002892, + "learning_rate": 0.0006, + "loss": 4.356374740600586, + "step": 6090 + }, + { + "epoch": 84.60087336244541, + "grad_norm": 0.014166025444865227, + "learning_rate": 0.0006, + "loss": 4.165982246398926, + "step": 6091 + }, + { + "epoch": 84.61484716157206, + "grad_norm": 0.012198803946375847, + "learning_rate": 0.0006, + "loss": 4.287566184997559, + "step": 6092 + }, + { + "epoch": 84.62882096069869, + "grad_norm": 0.013794238679111004, + "learning_rate": 0.0006, + "loss": 4.199882984161377, + "step": 6093 + }, + { + "epoch": 84.64279475982532, + "grad_norm": 0.014757642522454262, + "learning_rate": 0.0006, + "loss": 4.116573333740234, + "step": 6094 + }, + { + "epoch": 84.65676855895197, + "grad_norm": 0.014591689221560955, + "learning_rate": 0.0006, + "loss": 4.240896224975586, + "step": 6095 + }, + { + "epoch": 84.6707423580786, + "grad_norm": 0.015217579901218414, + "learning_rate": 0.0006, + "loss": 4.257142066955566, + "step": 6096 + }, + { + "epoch": 84.68471615720524, + "grad_norm": 0.015371371991932392, + "learning_rate": 0.0006, + "loss": 4.181037425994873, + "step": 6097 + }, + { + "epoch": 84.69868995633188, + "grad_norm": 0.014412983320653439, + "learning_rate": 0.0006, + "loss": 4.2207794189453125, + "step": 6098 + }, + { + "epoch": 84.71266375545852, + "grad_norm": 0.013773671351373196, + "learning_rate": 0.0006, + "loss": 4.211123466491699, + "step": 6099 + }, + { + "epoch": 84.72663755458515, + "grad_norm": 0.012838046066462994, + "learning_rate": 0.0006, + "loss": 4.175459384918213, + "step": 6100 + }, + { + "epoch": 84.74061135371178, + "grad_norm": 0.014094929210841656, + "learning_rate": 0.0006, + "loss": 4.1724958419799805, + "step": 6101 + }, + { + "epoch": 84.75458515283843, + "grad_norm": 0.014573262073099613, + "learning_rate": 0.0006, + "loss": 4.15885066986084, + "step": 6102 + }, + { + "epoch": 84.76855895196506, + "grad_norm": 0.011964819394052029, + "learning_rate": 0.0006, + "loss": 3.9804892539978027, + "step": 6103 + }, + { + "epoch": 84.78253275109171, + "grad_norm": 0.014140539802610874, + "learning_rate": 0.0006, + "loss": 4.203708648681641, + "step": 6104 + }, + { + "epoch": 84.79650655021834, + "grad_norm": 0.016618940979242325, + "learning_rate": 0.0006, + "loss": 4.147684097290039, + "step": 6105 + }, + { + "epoch": 84.81048034934499, + "grad_norm": 0.015094385482370853, + "learning_rate": 0.0006, + "loss": 4.392223834991455, + "step": 6106 + }, + { + "epoch": 84.82445414847162, + "grad_norm": 0.014386717230081558, + "learning_rate": 0.0006, + "loss": 4.253332138061523, + "step": 6107 + }, + { + "epoch": 84.83842794759825, + "grad_norm": 0.013877755962312222, + "learning_rate": 0.0006, + "loss": 4.075056076049805, + "step": 6108 + }, + { + "epoch": 84.8524017467249, + "grad_norm": 0.014062155969440937, + "learning_rate": 0.0006, + "loss": 4.072048664093018, + "step": 6109 + }, + { + "epoch": 84.86637554585153, + "grad_norm": 0.014445619657635689, + "learning_rate": 0.0006, + "loss": 4.151832580566406, + "step": 6110 + }, + { + "epoch": 84.88034934497817, + "grad_norm": 0.015116363763809204, + "learning_rate": 0.0006, + "loss": 4.088900566101074, + "step": 6111 + }, + { + "epoch": 84.8943231441048, + "grad_norm": 0.014425474219024181, + "learning_rate": 0.0006, + "loss": 4.200483322143555, + "step": 6112 + }, + { + "epoch": 84.90829694323143, + "grad_norm": 0.01449573878198862, + "learning_rate": 0.0006, + "loss": 4.239742279052734, + "step": 6113 + }, + { + "epoch": 84.92227074235808, + "grad_norm": 0.015698261559009552, + "learning_rate": 0.0006, + "loss": 4.197391510009766, + "step": 6114 + }, + { + "epoch": 84.93624454148471, + "grad_norm": 0.014545733109116554, + "learning_rate": 0.0006, + "loss": 4.035508155822754, + "step": 6115 + }, + { + "epoch": 84.95021834061136, + "grad_norm": 0.012796483002603054, + "learning_rate": 0.0006, + "loss": 4.195248126983643, + "step": 6116 + }, + { + "epoch": 84.96419213973799, + "grad_norm": 0.01264485064893961, + "learning_rate": 0.0006, + "loss": 4.158118724822998, + "step": 6117 + }, + { + "epoch": 84.97816593886463, + "grad_norm": 0.013695276342332363, + "learning_rate": 0.0006, + "loss": 4.255288124084473, + "step": 6118 + }, + { + "epoch": 84.99213973799127, + "grad_norm": 0.013459564186632633, + "learning_rate": 0.0006, + "loss": 4.204586505889893, + "step": 6119 + }, + { + "epoch": 85.0, + "grad_norm": 0.014312652871012688, + "learning_rate": 0.0006, + "loss": 4.167879104614258, + "step": 6120 + }, + { + "epoch": 85.0, + "eval_loss": 4.608152389526367, + "eval_runtime": 57.1903, + "eval_samples_per_second": 42.7, + "eval_steps_per_second": 1.346, + "step": 6120 + }, + { + "epoch": 85.01397379912663, + "grad_norm": 0.013918368145823479, + "learning_rate": 0.0006, + "loss": 4.129991054534912, + "step": 6121 + }, + { + "epoch": 85.02794759825328, + "grad_norm": 0.01604384370148182, + "learning_rate": 0.0006, + "loss": 4.233926773071289, + "step": 6122 + }, + { + "epoch": 85.04192139737991, + "grad_norm": 0.016565652564167976, + "learning_rate": 0.0006, + "loss": 4.129604339599609, + "step": 6123 + }, + { + "epoch": 85.05589519650655, + "grad_norm": 0.0149751678109169, + "learning_rate": 0.0006, + "loss": 4.165839195251465, + "step": 6124 + }, + { + "epoch": 85.06986899563319, + "grad_norm": 0.013812461867928505, + "learning_rate": 0.0006, + "loss": 4.153792858123779, + "step": 6125 + }, + { + "epoch": 85.08384279475983, + "grad_norm": 0.014432745985686779, + "learning_rate": 0.0006, + "loss": 4.170895576477051, + "step": 6126 + }, + { + "epoch": 85.09781659388646, + "grad_norm": 0.015208868309855461, + "learning_rate": 0.0006, + "loss": 4.10639762878418, + "step": 6127 + }, + { + "epoch": 85.1117903930131, + "grad_norm": 0.013209272176027298, + "learning_rate": 0.0006, + "loss": 4.200364112854004, + "step": 6128 + }, + { + "epoch": 85.12576419213974, + "grad_norm": 0.014807002618908882, + "learning_rate": 0.0006, + "loss": 4.172666549682617, + "step": 6129 + }, + { + "epoch": 85.13973799126637, + "grad_norm": 0.015662945806980133, + "learning_rate": 0.0006, + "loss": 4.145219326019287, + "step": 6130 + }, + { + "epoch": 85.15371179039302, + "grad_norm": 0.014598497189581394, + "learning_rate": 0.0006, + "loss": 4.275618553161621, + "step": 6131 + }, + { + "epoch": 85.16768558951965, + "grad_norm": 0.013778576627373695, + "learning_rate": 0.0006, + "loss": 4.204716682434082, + "step": 6132 + }, + { + "epoch": 85.18165938864628, + "grad_norm": 0.015762265771627426, + "learning_rate": 0.0006, + "loss": 4.145524024963379, + "step": 6133 + }, + { + "epoch": 85.19563318777293, + "grad_norm": 0.015579195693135262, + "learning_rate": 0.0006, + "loss": 4.150625228881836, + "step": 6134 + }, + { + "epoch": 85.20960698689956, + "grad_norm": 0.016641216352581978, + "learning_rate": 0.0006, + "loss": 4.141819953918457, + "step": 6135 + }, + { + "epoch": 85.2235807860262, + "grad_norm": 0.01632271707057953, + "learning_rate": 0.0006, + "loss": 4.030620574951172, + "step": 6136 + }, + { + "epoch": 85.23755458515284, + "grad_norm": 0.015564918518066406, + "learning_rate": 0.0006, + "loss": 4.147753715515137, + "step": 6137 + }, + { + "epoch": 85.25152838427948, + "grad_norm": 0.0161129217594862, + "learning_rate": 0.0006, + "loss": 4.128524303436279, + "step": 6138 + }, + { + "epoch": 85.26550218340611, + "grad_norm": 0.017766550183296204, + "learning_rate": 0.0006, + "loss": 4.134856224060059, + "step": 6139 + }, + { + "epoch": 85.27947598253274, + "grad_norm": 0.01869054324924946, + "learning_rate": 0.0006, + "loss": 4.231565475463867, + "step": 6140 + }, + { + "epoch": 85.29344978165939, + "grad_norm": 0.01917010359466076, + "learning_rate": 0.0006, + "loss": 4.21175479888916, + "step": 6141 + }, + { + "epoch": 85.30742358078602, + "grad_norm": 0.018116485327482224, + "learning_rate": 0.0006, + "loss": 4.122644424438477, + "step": 6142 + }, + { + "epoch": 85.32139737991267, + "grad_norm": 0.015949107706546783, + "learning_rate": 0.0006, + "loss": 4.135458946228027, + "step": 6143 + }, + { + "epoch": 85.3353711790393, + "grad_norm": 0.014783683232963085, + "learning_rate": 0.0006, + "loss": 4.118994235992432, + "step": 6144 + }, + { + "epoch": 85.34934497816595, + "grad_norm": 0.015406692400574684, + "learning_rate": 0.0006, + "loss": 4.157268524169922, + "step": 6145 + }, + { + "epoch": 85.36331877729258, + "grad_norm": 0.013893205672502518, + "learning_rate": 0.0006, + "loss": 4.056166648864746, + "step": 6146 + }, + { + "epoch": 85.37729257641921, + "grad_norm": 0.014380923472344875, + "learning_rate": 0.0006, + "loss": 4.206905364990234, + "step": 6147 + }, + { + "epoch": 85.39126637554585, + "grad_norm": 0.015669258311390877, + "learning_rate": 0.0006, + "loss": 4.213881492614746, + "step": 6148 + }, + { + "epoch": 85.40524017467249, + "grad_norm": 0.01553428452461958, + "learning_rate": 0.0006, + "loss": 4.1814374923706055, + "step": 6149 + }, + { + "epoch": 85.41921397379913, + "grad_norm": 0.01697700284421444, + "learning_rate": 0.0006, + "loss": 4.083939552307129, + "step": 6150 + }, + { + "epoch": 85.43318777292576, + "grad_norm": 0.016639070585370064, + "learning_rate": 0.0006, + "loss": 4.279311180114746, + "step": 6151 + }, + { + "epoch": 85.44716157205241, + "grad_norm": 0.016285281628370285, + "learning_rate": 0.0006, + "loss": 4.244845390319824, + "step": 6152 + }, + { + "epoch": 85.46113537117904, + "grad_norm": 0.016071317717432976, + "learning_rate": 0.0006, + "loss": 4.089814186096191, + "step": 6153 + }, + { + "epoch": 85.47510917030567, + "grad_norm": 0.01640418916940689, + "learning_rate": 0.0006, + "loss": 4.168088912963867, + "step": 6154 + }, + { + "epoch": 85.48908296943232, + "grad_norm": 0.014781414531171322, + "learning_rate": 0.0006, + "loss": 4.196092128753662, + "step": 6155 + }, + { + "epoch": 85.50305676855895, + "grad_norm": 0.013393213972449303, + "learning_rate": 0.0006, + "loss": 4.193429946899414, + "step": 6156 + }, + { + "epoch": 85.5170305676856, + "grad_norm": 0.01467406190931797, + "learning_rate": 0.0006, + "loss": 4.216714859008789, + "step": 6157 + }, + { + "epoch": 85.53100436681223, + "grad_norm": 0.014950765296816826, + "learning_rate": 0.0006, + "loss": 4.093854904174805, + "step": 6158 + }, + { + "epoch": 85.54497816593886, + "grad_norm": 0.014655301347374916, + "learning_rate": 0.0006, + "loss": 4.09483528137207, + "step": 6159 + }, + { + "epoch": 85.5589519650655, + "grad_norm": 0.01425520796328783, + "learning_rate": 0.0006, + "loss": 4.090396881103516, + "step": 6160 + }, + { + "epoch": 85.57292576419214, + "grad_norm": 0.01403278298676014, + "learning_rate": 0.0006, + "loss": 4.088724136352539, + "step": 6161 + }, + { + "epoch": 85.58689956331878, + "grad_norm": 0.01364967692643404, + "learning_rate": 0.0006, + "loss": 4.095558166503906, + "step": 6162 + }, + { + "epoch": 85.60087336244541, + "grad_norm": 0.013085267506539822, + "learning_rate": 0.0006, + "loss": 4.163962364196777, + "step": 6163 + }, + { + "epoch": 85.61484716157206, + "grad_norm": 0.0136906448751688, + "learning_rate": 0.0006, + "loss": 4.197134017944336, + "step": 6164 + }, + { + "epoch": 85.62882096069869, + "grad_norm": 0.013472451828420162, + "learning_rate": 0.0006, + "loss": 4.1578369140625, + "step": 6165 + }, + { + "epoch": 85.64279475982532, + "grad_norm": 0.012329773046076298, + "learning_rate": 0.0006, + "loss": 4.231907844543457, + "step": 6166 + }, + { + "epoch": 85.65676855895197, + "grad_norm": 0.011977200396358967, + "learning_rate": 0.0006, + "loss": 4.1552958488464355, + "step": 6167 + }, + { + "epoch": 85.6707423580786, + "grad_norm": 0.013049956411123276, + "learning_rate": 0.0006, + "loss": 4.17852783203125, + "step": 6168 + }, + { + "epoch": 85.68471615720524, + "grad_norm": 0.011440741829574108, + "learning_rate": 0.0006, + "loss": 4.144646644592285, + "step": 6169 + }, + { + "epoch": 85.69868995633188, + "grad_norm": 0.012472936883568764, + "learning_rate": 0.0006, + "loss": 4.2306413650512695, + "step": 6170 + }, + { + "epoch": 85.71266375545852, + "grad_norm": 0.013015356846153736, + "learning_rate": 0.0006, + "loss": 4.193439960479736, + "step": 6171 + }, + { + "epoch": 85.72663755458515, + "grad_norm": 0.013832319527864456, + "learning_rate": 0.0006, + "loss": 4.18073034286499, + "step": 6172 + }, + { + "epoch": 85.74061135371178, + "grad_norm": 0.014613982290029526, + "learning_rate": 0.0006, + "loss": 4.223531246185303, + "step": 6173 + }, + { + "epoch": 85.75458515283843, + "grad_norm": 0.014209083281457424, + "learning_rate": 0.0006, + "loss": 4.142666816711426, + "step": 6174 + }, + { + "epoch": 85.76855895196506, + "grad_norm": 0.013396669179201126, + "learning_rate": 0.0006, + "loss": 4.185786247253418, + "step": 6175 + }, + { + "epoch": 85.78253275109171, + "grad_norm": 0.01364929135888815, + "learning_rate": 0.0006, + "loss": 4.125226974487305, + "step": 6176 + }, + { + "epoch": 85.79650655021834, + "grad_norm": 0.012561003677546978, + "learning_rate": 0.0006, + "loss": 4.084519386291504, + "step": 6177 + }, + { + "epoch": 85.81048034934499, + "grad_norm": 0.013508481904864311, + "learning_rate": 0.0006, + "loss": 4.193173885345459, + "step": 6178 + }, + { + "epoch": 85.82445414847162, + "grad_norm": 0.014154613018035889, + "learning_rate": 0.0006, + "loss": 4.257272720336914, + "step": 6179 + }, + { + "epoch": 85.83842794759825, + "grad_norm": 0.012478531338274479, + "learning_rate": 0.0006, + "loss": 4.161380767822266, + "step": 6180 + }, + { + "epoch": 85.8524017467249, + "grad_norm": 0.013468451797962189, + "learning_rate": 0.0006, + "loss": 4.121395111083984, + "step": 6181 + }, + { + "epoch": 85.86637554585153, + "grad_norm": 0.015155194327235222, + "learning_rate": 0.0006, + "loss": 4.242815971374512, + "step": 6182 + }, + { + "epoch": 85.88034934497817, + "grad_norm": 0.016223272308707237, + "learning_rate": 0.0006, + "loss": 4.176027297973633, + "step": 6183 + }, + { + "epoch": 85.8943231441048, + "grad_norm": 0.013741295784711838, + "learning_rate": 0.0006, + "loss": 4.131257057189941, + "step": 6184 + }, + { + "epoch": 85.90829694323143, + "grad_norm": 0.014982067979872227, + "learning_rate": 0.0006, + "loss": 4.237896919250488, + "step": 6185 + }, + { + "epoch": 85.92227074235808, + "grad_norm": 0.018197322264313698, + "learning_rate": 0.0006, + "loss": 4.154962539672852, + "step": 6186 + }, + { + "epoch": 85.93624454148471, + "grad_norm": 0.01971200481057167, + "learning_rate": 0.0006, + "loss": 4.16819429397583, + "step": 6187 + }, + { + "epoch": 85.95021834061136, + "grad_norm": 0.018008720129728317, + "learning_rate": 0.0006, + "loss": 4.146164417266846, + "step": 6188 + }, + { + "epoch": 85.96419213973799, + "grad_norm": 0.015859264880418777, + "learning_rate": 0.0006, + "loss": 4.208099365234375, + "step": 6189 + }, + { + "epoch": 85.97816593886463, + "grad_norm": 0.013846252113580704, + "learning_rate": 0.0006, + "loss": 4.1881184577941895, + "step": 6190 + }, + { + "epoch": 85.99213973799127, + "grad_norm": 0.018641939386725426, + "learning_rate": 0.0006, + "loss": 4.127632141113281, + "step": 6191 + }, + { + "epoch": 86.0, + "grad_norm": 0.02385227382183075, + "learning_rate": 0.0006, + "loss": 4.281900405883789, + "step": 6192 + }, + { + "epoch": 86.0, + "eval_loss": 4.625458717346191, + "eval_runtime": 57.0634, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6192 + }, + { + "epoch": 86.01397379912663, + "grad_norm": 0.019032124429941177, + "learning_rate": 0.0006, + "loss": 4.2163190841674805, + "step": 6193 + }, + { + "epoch": 86.02794759825328, + "grad_norm": 0.018049925565719604, + "learning_rate": 0.0006, + "loss": 4.1534037590026855, + "step": 6194 + }, + { + "epoch": 86.04192139737991, + "grad_norm": 0.019116360694169998, + "learning_rate": 0.0006, + "loss": 4.212411403656006, + "step": 6195 + }, + { + "epoch": 86.05589519650655, + "grad_norm": 0.018967336043715477, + "learning_rate": 0.0006, + "loss": 4.190918922424316, + "step": 6196 + }, + { + "epoch": 86.06986899563319, + "grad_norm": 0.01945611648261547, + "learning_rate": 0.0006, + "loss": 4.18459415435791, + "step": 6197 + }, + { + "epoch": 86.08384279475983, + "grad_norm": 0.018104076385498047, + "learning_rate": 0.0006, + "loss": 4.0942816734313965, + "step": 6198 + }, + { + "epoch": 86.09781659388646, + "grad_norm": 0.01815992221236229, + "learning_rate": 0.0006, + "loss": 4.011547565460205, + "step": 6199 + }, + { + "epoch": 86.1117903930131, + "grad_norm": 0.017574431374669075, + "learning_rate": 0.0006, + "loss": 4.051593780517578, + "step": 6200 + }, + { + "epoch": 86.12576419213974, + "grad_norm": 0.01516508124768734, + "learning_rate": 0.0006, + "loss": 4.239673614501953, + "step": 6201 + }, + { + "epoch": 86.13973799126637, + "grad_norm": 0.015689702704548836, + "learning_rate": 0.0006, + "loss": 4.1085734367370605, + "step": 6202 + }, + { + "epoch": 86.15371179039302, + "grad_norm": 0.01596900261938572, + "learning_rate": 0.0006, + "loss": 4.1509599685668945, + "step": 6203 + }, + { + "epoch": 86.16768558951965, + "grad_norm": 0.01550096832215786, + "learning_rate": 0.0006, + "loss": 4.160238265991211, + "step": 6204 + }, + { + "epoch": 86.18165938864628, + "grad_norm": 0.014601529575884342, + "learning_rate": 0.0006, + "loss": 4.029219150543213, + "step": 6205 + }, + { + "epoch": 86.19563318777293, + "grad_norm": 0.014006740413606167, + "learning_rate": 0.0006, + "loss": 4.117501258850098, + "step": 6206 + }, + { + "epoch": 86.20960698689956, + "grad_norm": 0.013191957026720047, + "learning_rate": 0.0006, + "loss": 4.252002716064453, + "step": 6207 + }, + { + "epoch": 86.2235807860262, + "grad_norm": 0.013031814247369766, + "learning_rate": 0.0006, + "loss": 4.209517478942871, + "step": 6208 + }, + { + "epoch": 86.23755458515284, + "grad_norm": 0.013605714775621891, + "learning_rate": 0.0006, + "loss": 4.176309585571289, + "step": 6209 + }, + { + "epoch": 86.25152838427948, + "grad_norm": 0.012753087095916271, + "learning_rate": 0.0006, + "loss": 4.097990036010742, + "step": 6210 + }, + { + "epoch": 86.26550218340611, + "grad_norm": 0.012940628454089165, + "learning_rate": 0.0006, + "loss": 4.085589408874512, + "step": 6211 + }, + { + "epoch": 86.27947598253274, + "grad_norm": 0.014286619611084461, + "learning_rate": 0.0006, + "loss": 4.199261665344238, + "step": 6212 + }, + { + "epoch": 86.29344978165939, + "grad_norm": 0.014643525704741478, + "learning_rate": 0.0006, + "loss": 4.23718786239624, + "step": 6213 + }, + { + "epoch": 86.30742358078602, + "grad_norm": 0.015182147733867168, + "learning_rate": 0.0006, + "loss": 4.219855308532715, + "step": 6214 + }, + { + "epoch": 86.32139737991267, + "grad_norm": 0.01530102826654911, + "learning_rate": 0.0006, + "loss": 4.190094947814941, + "step": 6215 + }, + { + "epoch": 86.3353711790393, + "grad_norm": 0.015492603182792664, + "learning_rate": 0.0006, + "loss": 4.153780460357666, + "step": 6216 + }, + { + "epoch": 86.34934497816595, + "grad_norm": 0.015339907258749008, + "learning_rate": 0.0006, + "loss": 4.182865619659424, + "step": 6217 + }, + { + "epoch": 86.36331877729258, + "grad_norm": 0.015576880425214767, + "learning_rate": 0.0006, + "loss": 4.196417331695557, + "step": 6218 + }, + { + "epoch": 86.37729257641921, + "grad_norm": 0.014805737882852554, + "learning_rate": 0.0006, + "loss": 4.265018463134766, + "step": 6219 + }, + { + "epoch": 86.39126637554585, + "grad_norm": 0.01597665622830391, + "learning_rate": 0.0006, + "loss": 4.2864837646484375, + "step": 6220 + }, + { + "epoch": 86.40524017467249, + "grad_norm": 0.01693643443286419, + "learning_rate": 0.0006, + "loss": 4.118192672729492, + "step": 6221 + }, + { + "epoch": 86.41921397379913, + "grad_norm": 0.016648951917886734, + "learning_rate": 0.0006, + "loss": 4.23712682723999, + "step": 6222 + }, + { + "epoch": 86.43318777292576, + "grad_norm": 0.014809766784310341, + "learning_rate": 0.0006, + "loss": 4.251391410827637, + "step": 6223 + }, + { + "epoch": 86.44716157205241, + "grad_norm": 0.014333373866975307, + "learning_rate": 0.0006, + "loss": 4.17243766784668, + "step": 6224 + }, + { + "epoch": 86.46113537117904, + "grad_norm": 0.013912711292505264, + "learning_rate": 0.0006, + "loss": 4.250836372375488, + "step": 6225 + }, + { + "epoch": 86.47510917030567, + "grad_norm": 0.014821341261267662, + "learning_rate": 0.0006, + "loss": 4.073995590209961, + "step": 6226 + }, + { + "epoch": 86.48908296943232, + "grad_norm": 0.016634922474622726, + "learning_rate": 0.0006, + "loss": 4.155867576599121, + "step": 6227 + }, + { + "epoch": 86.50305676855895, + "grad_norm": 0.01748695969581604, + "learning_rate": 0.0006, + "loss": 4.139344692230225, + "step": 6228 + }, + { + "epoch": 86.5170305676856, + "grad_norm": 0.01569363661110401, + "learning_rate": 0.0006, + "loss": 4.23923397064209, + "step": 6229 + }, + { + "epoch": 86.53100436681223, + "grad_norm": 0.014037121087312698, + "learning_rate": 0.0006, + "loss": 4.200334072113037, + "step": 6230 + }, + { + "epoch": 86.54497816593886, + "grad_norm": 0.01401279866695404, + "learning_rate": 0.0006, + "loss": 4.035246849060059, + "step": 6231 + }, + { + "epoch": 86.5589519650655, + "grad_norm": 0.014603731222450733, + "learning_rate": 0.0006, + "loss": 4.141385555267334, + "step": 6232 + }, + { + "epoch": 86.57292576419214, + "grad_norm": 0.015113704837858677, + "learning_rate": 0.0006, + "loss": 4.130824089050293, + "step": 6233 + }, + { + "epoch": 86.58689956331878, + "grad_norm": 0.01588696800172329, + "learning_rate": 0.0006, + "loss": 4.05495023727417, + "step": 6234 + }, + { + "epoch": 86.60087336244541, + "grad_norm": 0.01593640074133873, + "learning_rate": 0.0006, + "loss": 4.22935152053833, + "step": 6235 + }, + { + "epoch": 86.61484716157206, + "grad_norm": 0.014292445965111256, + "learning_rate": 0.0006, + "loss": 4.309839248657227, + "step": 6236 + }, + { + "epoch": 86.62882096069869, + "grad_norm": 0.015471610240638256, + "learning_rate": 0.0006, + "loss": 4.268278121948242, + "step": 6237 + }, + { + "epoch": 86.64279475982532, + "grad_norm": 0.016643544659018517, + "learning_rate": 0.0006, + "loss": 4.135109901428223, + "step": 6238 + }, + { + "epoch": 86.65676855895197, + "grad_norm": 0.014733029529452324, + "learning_rate": 0.0006, + "loss": 4.060197830200195, + "step": 6239 + }, + { + "epoch": 86.6707423580786, + "grad_norm": 0.014869216829538345, + "learning_rate": 0.0006, + "loss": 4.126917362213135, + "step": 6240 + }, + { + "epoch": 86.68471615720524, + "grad_norm": 0.01608203910291195, + "learning_rate": 0.0006, + "loss": 4.183716773986816, + "step": 6241 + }, + { + "epoch": 86.69868995633188, + "grad_norm": 0.016178175806999207, + "learning_rate": 0.0006, + "loss": 4.118875503540039, + "step": 6242 + }, + { + "epoch": 86.71266375545852, + "grad_norm": 0.014576597139239311, + "learning_rate": 0.0006, + "loss": 4.261541366577148, + "step": 6243 + }, + { + "epoch": 86.72663755458515, + "grad_norm": 0.014466420747339725, + "learning_rate": 0.0006, + "loss": 4.307284355163574, + "step": 6244 + }, + { + "epoch": 86.74061135371178, + "grad_norm": 0.014042644761502743, + "learning_rate": 0.0006, + "loss": 4.2729997634887695, + "step": 6245 + }, + { + "epoch": 86.75458515283843, + "grad_norm": 0.014589524827897549, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6246 + }, + { + "epoch": 86.76855895196506, + "grad_norm": 0.018765483051538467, + "learning_rate": 0.0006, + "loss": 4.264769554138184, + "step": 6247 + }, + { + "epoch": 86.78253275109171, + "grad_norm": 0.019397731870412827, + "learning_rate": 0.0006, + "loss": 4.183379173278809, + "step": 6248 + }, + { + "epoch": 86.79650655021834, + "grad_norm": 0.017525402829051018, + "learning_rate": 0.0006, + "loss": 4.1494245529174805, + "step": 6249 + }, + { + "epoch": 86.81048034934499, + "grad_norm": 0.016310200095176697, + "learning_rate": 0.0006, + "loss": 4.149104118347168, + "step": 6250 + }, + { + "epoch": 86.82445414847162, + "grad_norm": 0.01506423857063055, + "learning_rate": 0.0006, + "loss": 4.22091817855835, + "step": 6251 + }, + { + "epoch": 86.83842794759825, + "grad_norm": 0.014200284145772457, + "learning_rate": 0.0006, + "loss": 4.177569389343262, + "step": 6252 + }, + { + "epoch": 86.8524017467249, + "grad_norm": 0.015016552060842514, + "learning_rate": 0.0006, + "loss": 4.089859962463379, + "step": 6253 + }, + { + "epoch": 86.86637554585153, + "grad_norm": 0.016054971143603325, + "learning_rate": 0.0006, + "loss": 4.157166004180908, + "step": 6254 + }, + { + "epoch": 86.88034934497817, + "grad_norm": 0.014388521201908588, + "learning_rate": 0.0006, + "loss": 4.1437578201293945, + "step": 6255 + }, + { + "epoch": 86.8943231441048, + "grad_norm": 0.015705369412899017, + "learning_rate": 0.0006, + "loss": 4.045689582824707, + "step": 6256 + }, + { + "epoch": 86.90829694323143, + "grad_norm": 0.014192395843565464, + "learning_rate": 0.0006, + "loss": 4.098793029785156, + "step": 6257 + }, + { + "epoch": 86.92227074235808, + "grad_norm": 0.013751034624874592, + "learning_rate": 0.0006, + "loss": 3.989920139312744, + "step": 6258 + }, + { + "epoch": 86.93624454148471, + "grad_norm": 0.01582193188369274, + "learning_rate": 0.0006, + "loss": 4.283449172973633, + "step": 6259 + }, + { + "epoch": 86.95021834061136, + "grad_norm": 0.01476013008505106, + "learning_rate": 0.0006, + "loss": 4.221401691436768, + "step": 6260 + }, + { + "epoch": 86.96419213973799, + "grad_norm": 0.01346337515860796, + "learning_rate": 0.0006, + "loss": 4.05896520614624, + "step": 6261 + }, + { + "epoch": 86.97816593886463, + "grad_norm": 0.012899307534098625, + "learning_rate": 0.0006, + "loss": 4.159560203552246, + "step": 6262 + }, + { + "epoch": 86.99213973799127, + "grad_norm": 0.013150574639439583, + "learning_rate": 0.0006, + "loss": 4.217330455780029, + "step": 6263 + }, + { + "epoch": 87.0, + "grad_norm": 0.014940370805561543, + "learning_rate": 0.0006, + "loss": 4.1611833572387695, + "step": 6264 + }, + { + "epoch": 87.0, + "eval_loss": 4.666043758392334, + "eval_runtime": 56.5978, + "eval_samples_per_second": 43.147, + "eval_steps_per_second": 1.36, + "step": 6264 + }, + { + "epoch": 87.01397379912663, + "grad_norm": 0.013721180148422718, + "learning_rate": 0.0006, + "loss": 4.142454147338867, + "step": 6265 + }, + { + "epoch": 87.02794759825328, + "grad_norm": 0.014805255457758904, + "learning_rate": 0.0006, + "loss": 4.121677875518799, + "step": 6266 + }, + { + "epoch": 87.04192139737991, + "grad_norm": 0.015576514415442944, + "learning_rate": 0.0006, + "loss": 4.1188435554504395, + "step": 6267 + }, + { + "epoch": 87.05589519650655, + "grad_norm": 0.01605502888560295, + "learning_rate": 0.0006, + "loss": 4.255428314208984, + "step": 6268 + }, + { + "epoch": 87.06986899563319, + "grad_norm": 0.014933343045413494, + "learning_rate": 0.0006, + "loss": 4.122596740722656, + "step": 6269 + }, + { + "epoch": 87.08384279475983, + "grad_norm": 0.014630894176661968, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 6270 + }, + { + "epoch": 87.09781659388646, + "grad_norm": 0.014211163856089115, + "learning_rate": 0.0006, + "loss": 4.058030128479004, + "step": 6271 + }, + { + "epoch": 87.1117903930131, + "grad_norm": 0.014060168527066708, + "learning_rate": 0.0006, + "loss": 4.146586894989014, + "step": 6272 + }, + { + "epoch": 87.12576419213974, + "grad_norm": 0.014740100130438805, + "learning_rate": 0.0006, + "loss": 4.077739715576172, + "step": 6273 + }, + { + "epoch": 87.13973799126637, + "grad_norm": 0.013970664702355862, + "learning_rate": 0.0006, + "loss": 4.038252830505371, + "step": 6274 + }, + { + "epoch": 87.15371179039302, + "grad_norm": 0.014399951323866844, + "learning_rate": 0.0006, + "loss": 4.143700122833252, + "step": 6275 + }, + { + "epoch": 87.16768558951965, + "grad_norm": 0.01496410183608532, + "learning_rate": 0.0006, + "loss": 4.105609893798828, + "step": 6276 + }, + { + "epoch": 87.18165938864628, + "grad_norm": 0.014287355355918407, + "learning_rate": 0.0006, + "loss": 4.173645973205566, + "step": 6277 + }, + { + "epoch": 87.19563318777293, + "grad_norm": 0.0141292130574584, + "learning_rate": 0.0006, + "loss": 4.068511486053467, + "step": 6278 + }, + { + "epoch": 87.20960698689956, + "grad_norm": 0.01393523346632719, + "learning_rate": 0.0006, + "loss": 4.060687065124512, + "step": 6279 + }, + { + "epoch": 87.2235807860262, + "grad_norm": 0.01431021187454462, + "learning_rate": 0.0006, + "loss": 4.219209671020508, + "step": 6280 + }, + { + "epoch": 87.23755458515284, + "grad_norm": 0.014856027439236641, + "learning_rate": 0.0006, + "loss": 4.146749973297119, + "step": 6281 + }, + { + "epoch": 87.25152838427948, + "grad_norm": 0.013677539303898811, + "learning_rate": 0.0006, + "loss": 3.9859797954559326, + "step": 6282 + }, + { + "epoch": 87.26550218340611, + "grad_norm": 0.014078967273235321, + "learning_rate": 0.0006, + "loss": 4.129668235778809, + "step": 6283 + }, + { + "epoch": 87.27947598253274, + "grad_norm": 0.015564394183456898, + "learning_rate": 0.0006, + "loss": 4.0007829666137695, + "step": 6284 + }, + { + "epoch": 87.29344978165939, + "grad_norm": 0.015077749267220497, + "learning_rate": 0.0006, + "loss": 4.112264156341553, + "step": 6285 + }, + { + "epoch": 87.30742358078602, + "grad_norm": 0.015341218560934067, + "learning_rate": 0.0006, + "loss": 4.112249374389648, + "step": 6286 + }, + { + "epoch": 87.32139737991267, + "grad_norm": 0.016399366781115532, + "learning_rate": 0.0006, + "loss": 4.176196098327637, + "step": 6287 + }, + { + "epoch": 87.3353711790393, + "grad_norm": 0.01599763333797455, + "learning_rate": 0.0006, + "loss": 4.163441181182861, + "step": 6288 + }, + { + "epoch": 87.34934497816595, + "grad_norm": 0.015439776703715324, + "learning_rate": 0.0006, + "loss": 4.11902379989624, + "step": 6289 + }, + { + "epoch": 87.36331877729258, + "grad_norm": 0.015909943729639053, + "learning_rate": 0.0006, + "loss": 4.16621732711792, + "step": 6290 + }, + { + "epoch": 87.37729257641921, + "grad_norm": 0.016710078343749046, + "learning_rate": 0.0006, + "loss": 4.044960975646973, + "step": 6291 + }, + { + "epoch": 87.39126637554585, + "grad_norm": 0.015883933752775192, + "learning_rate": 0.0006, + "loss": 4.065184593200684, + "step": 6292 + }, + { + "epoch": 87.40524017467249, + "grad_norm": 0.013912747614085674, + "learning_rate": 0.0006, + "loss": 4.102450370788574, + "step": 6293 + }, + { + "epoch": 87.41921397379913, + "grad_norm": 0.013170420192182064, + "learning_rate": 0.0006, + "loss": 4.234364032745361, + "step": 6294 + }, + { + "epoch": 87.43318777292576, + "grad_norm": 0.014043302275240421, + "learning_rate": 0.0006, + "loss": 4.129565238952637, + "step": 6295 + }, + { + "epoch": 87.44716157205241, + "grad_norm": 0.015010139904916286, + "learning_rate": 0.0006, + "loss": 4.166678428649902, + "step": 6296 + }, + { + "epoch": 87.46113537117904, + "grad_norm": 0.014721784740686417, + "learning_rate": 0.0006, + "loss": 4.182618618011475, + "step": 6297 + }, + { + "epoch": 87.47510917030567, + "grad_norm": 0.015713468194007874, + "learning_rate": 0.0006, + "loss": 4.103855133056641, + "step": 6298 + }, + { + "epoch": 87.48908296943232, + "grad_norm": 0.017204247415065765, + "learning_rate": 0.0006, + "loss": 4.140937328338623, + "step": 6299 + }, + { + "epoch": 87.50305676855895, + "grad_norm": 0.016856668516993523, + "learning_rate": 0.0006, + "loss": 4.147859573364258, + "step": 6300 + }, + { + "epoch": 87.5170305676856, + "grad_norm": 0.017205018550157547, + "learning_rate": 0.0006, + "loss": 4.1667327880859375, + "step": 6301 + }, + { + "epoch": 87.53100436681223, + "grad_norm": 0.014821539632976055, + "learning_rate": 0.0006, + "loss": 4.089935302734375, + "step": 6302 + }, + { + "epoch": 87.54497816593886, + "grad_norm": 0.01446484588086605, + "learning_rate": 0.0006, + "loss": 4.243657112121582, + "step": 6303 + }, + { + "epoch": 87.5589519650655, + "grad_norm": 0.014374484308063984, + "learning_rate": 0.0006, + "loss": 4.14356803894043, + "step": 6304 + }, + { + "epoch": 87.57292576419214, + "grad_norm": 0.01420109998434782, + "learning_rate": 0.0006, + "loss": 4.1273651123046875, + "step": 6305 + }, + { + "epoch": 87.58689956331878, + "grad_norm": 0.014810388907790184, + "learning_rate": 0.0006, + "loss": 4.222290515899658, + "step": 6306 + }, + { + "epoch": 87.60087336244541, + "grad_norm": 0.015826983377337456, + "learning_rate": 0.0006, + "loss": 4.137996196746826, + "step": 6307 + }, + { + "epoch": 87.61484716157206, + "grad_norm": 0.014457454904913902, + "learning_rate": 0.0006, + "loss": 4.0780744552612305, + "step": 6308 + }, + { + "epoch": 87.62882096069869, + "grad_norm": 0.014458604156970978, + "learning_rate": 0.0006, + "loss": 4.1287922859191895, + "step": 6309 + }, + { + "epoch": 87.64279475982532, + "grad_norm": 0.015282071195542812, + "learning_rate": 0.0006, + "loss": 4.189508438110352, + "step": 6310 + }, + { + "epoch": 87.65676855895197, + "grad_norm": 0.017536642029881477, + "learning_rate": 0.0006, + "loss": 4.070173263549805, + "step": 6311 + }, + { + "epoch": 87.6707423580786, + "grad_norm": 0.017303580418229103, + "learning_rate": 0.0006, + "loss": 4.119133949279785, + "step": 6312 + }, + { + "epoch": 87.68471615720524, + "grad_norm": 0.016148580238223076, + "learning_rate": 0.0006, + "loss": 3.994354486465454, + "step": 6313 + }, + { + "epoch": 87.69868995633188, + "grad_norm": 0.01706942543387413, + "learning_rate": 0.0006, + "loss": 4.183113098144531, + "step": 6314 + }, + { + "epoch": 87.71266375545852, + "grad_norm": 0.016971010714769363, + "learning_rate": 0.0006, + "loss": 4.104086875915527, + "step": 6315 + }, + { + "epoch": 87.72663755458515, + "grad_norm": 0.015516847372055054, + "learning_rate": 0.0006, + "loss": 4.2409772872924805, + "step": 6316 + }, + { + "epoch": 87.74061135371178, + "grad_norm": 0.015352269634604454, + "learning_rate": 0.0006, + "loss": 4.026427268981934, + "step": 6317 + }, + { + "epoch": 87.75458515283843, + "grad_norm": 0.017702415585517883, + "learning_rate": 0.0006, + "loss": 4.177840232849121, + "step": 6318 + }, + { + "epoch": 87.76855895196506, + "grad_norm": 0.01629541628062725, + "learning_rate": 0.0006, + "loss": 4.229039192199707, + "step": 6319 + }, + { + "epoch": 87.78253275109171, + "grad_norm": 0.01503711100667715, + "learning_rate": 0.0006, + "loss": 4.214350700378418, + "step": 6320 + }, + { + "epoch": 87.79650655021834, + "grad_norm": 0.014913883060216904, + "learning_rate": 0.0006, + "loss": 4.168880939483643, + "step": 6321 + }, + { + "epoch": 87.81048034934499, + "grad_norm": 0.014513909816741943, + "learning_rate": 0.0006, + "loss": 4.162267684936523, + "step": 6322 + }, + { + "epoch": 87.82445414847162, + "grad_norm": 0.014635873027145863, + "learning_rate": 0.0006, + "loss": 4.220067977905273, + "step": 6323 + }, + { + "epoch": 87.83842794759825, + "grad_norm": 0.017102796584367752, + "learning_rate": 0.0006, + "loss": 4.1493377685546875, + "step": 6324 + }, + { + "epoch": 87.8524017467249, + "grad_norm": 0.015533884055912495, + "learning_rate": 0.0006, + "loss": 4.0646209716796875, + "step": 6325 + }, + { + "epoch": 87.86637554585153, + "grad_norm": 0.014617550186812878, + "learning_rate": 0.0006, + "loss": 4.1390533447265625, + "step": 6326 + }, + { + "epoch": 87.88034934497817, + "grad_norm": 0.014142967760562897, + "learning_rate": 0.0006, + "loss": 4.133934020996094, + "step": 6327 + }, + { + "epoch": 87.8943231441048, + "grad_norm": 0.013182433322072029, + "learning_rate": 0.0006, + "loss": 4.125327110290527, + "step": 6328 + }, + { + "epoch": 87.90829694323143, + "grad_norm": 0.013928229920566082, + "learning_rate": 0.0006, + "loss": 4.142777442932129, + "step": 6329 + }, + { + "epoch": 87.92227074235808, + "grad_norm": 0.015337013639509678, + "learning_rate": 0.0006, + "loss": 4.140210151672363, + "step": 6330 + }, + { + "epoch": 87.93624454148471, + "grad_norm": 0.014873536303639412, + "learning_rate": 0.0006, + "loss": 4.102636337280273, + "step": 6331 + }, + { + "epoch": 87.95021834061136, + "grad_norm": 0.01387431938201189, + "learning_rate": 0.0006, + "loss": 4.149059295654297, + "step": 6332 + }, + { + "epoch": 87.96419213973799, + "grad_norm": 0.013234943151473999, + "learning_rate": 0.0006, + "loss": 4.1564459800720215, + "step": 6333 + }, + { + "epoch": 87.97816593886463, + "grad_norm": 0.013781017623841763, + "learning_rate": 0.0006, + "loss": 4.189839839935303, + "step": 6334 + }, + { + "epoch": 87.99213973799127, + "grad_norm": 0.013733026571571827, + "learning_rate": 0.0006, + "loss": 4.29277229309082, + "step": 6335 + }, + { + "epoch": 88.0, + "grad_norm": 0.014752327464520931, + "learning_rate": 0.0006, + "loss": 4.163867950439453, + "step": 6336 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.701259757972357e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6336/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6336/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6336/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6624/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6624/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f487010d7fbcb1b812acb3764df63a019459ab80 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49a736af76c89dc83131d9cfcbb9f97ed70927d70d2fe82dbda55116bc87ddaf +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6624/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0d2cba0c5c5c417ed400c491752c71740d1963da --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e175ac6cb4d3df111ed04fca76ecb917ea42b452481a30c602abbc84283e744e +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..f7893f89b98ec10f03213b6e8beac3fa8d5e4926 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6d5bec1a5bb83a83aa97513fe583f8fc14f4840be26666e8cd7efa07536abe1 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..33aa04c3b93e992e98a21fd16e67f87011843073 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb8e96dd1435044d27f9f8cbcf84f8dd6c8f0f0749cf034035894986b2419766 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..8fff922a7792a11b3253e8dd288ca2c009b12152 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff27d8e23cd1e341ec2c1a3a22e887cd6b3e72b202f7c20aac714e64b8ad1825 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..9434f1eaa62b162a1f26d844c9cefc8483a39897 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4af1228878c5edb039eb391166ff4f3d6973c92e66fb7cd96e3fb2e64d0a2b3e +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6624/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1fb4fbb9cb8eafb61c9d89947d329fb67f4dd4e6 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aeab2b26cd983e17d83b05e83f37ef1706928049bbb8ff0e989945b479e25df2 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6624/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..30d8cf60cffd4ada8c2c7a0e02e68accd0abe0d1 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/trainer_state.json @@ -0,0 +1,47130 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 92.0, + "eval_steps": 500, + "global_step": 6624, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + }, + { + "epoch": 76.0, + "eval_loss": 4.689640998840332, + "eval_runtime": 56.9095, + "eval_samples_per_second": 42.91, + "eval_steps_per_second": 1.353, + "step": 5472 + }, + { + "epoch": 76.01397379912663, + "grad_norm": 0.015351055189967155, + "learning_rate": 0.0006, + "loss": 4.348268985748291, + "step": 5473 + }, + { + "epoch": 76.02794759825328, + "grad_norm": 0.017164217308163643, + "learning_rate": 0.0006, + "loss": 4.244204998016357, + "step": 5474 + }, + { + "epoch": 76.04192139737991, + "grad_norm": 0.021624932065606117, + "learning_rate": 0.0006, + "loss": 4.311168670654297, + "step": 5475 + }, + { + "epoch": 76.05589519650655, + "grad_norm": 0.028712468221783638, + "learning_rate": 0.0006, + "loss": 4.322287082672119, + "step": 5476 + }, + { + "epoch": 76.06986899563319, + "grad_norm": 0.027915487065911293, + "learning_rate": 0.0006, + "loss": 4.381207466125488, + "step": 5477 + }, + { + "epoch": 76.08384279475983, + "grad_norm": 0.017955303192138672, + "learning_rate": 0.0006, + "loss": 4.2635016441345215, + "step": 5478 + }, + { + "epoch": 76.09781659388646, + "grad_norm": 0.019113514572381973, + "learning_rate": 0.0006, + "loss": 4.399471282958984, + "step": 5479 + }, + { + "epoch": 76.1117903930131, + "grad_norm": 0.017502542585134506, + "learning_rate": 0.0006, + "loss": 4.321204662322998, + "step": 5480 + }, + { + "epoch": 76.12576419213974, + "grad_norm": 0.01852525770664215, + "learning_rate": 0.0006, + "loss": 4.217884063720703, + "step": 5481 + }, + { + "epoch": 76.13973799126637, + "grad_norm": 0.019836828112602234, + "learning_rate": 0.0006, + "loss": 4.278532028198242, + "step": 5482 + }, + { + "epoch": 76.15371179039302, + "grad_norm": 0.02101975679397583, + "learning_rate": 0.0006, + "loss": 4.246720314025879, + "step": 5483 + }, + { + "epoch": 76.16768558951965, + "grad_norm": 0.021202990785241127, + "learning_rate": 0.0006, + "loss": 4.517517566680908, + "step": 5484 + }, + { + "epoch": 76.18165938864628, + "grad_norm": 0.021428769454360008, + "learning_rate": 0.0006, + "loss": 4.3373517990112305, + "step": 5485 + }, + { + "epoch": 76.19563318777293, + "grad_norm": 0.018263190984725952, + "learning_rate": 0.0006, + "loss": 4.402856826782227, + "step": 5486 + }, + { + "epoch": 76.20960698689956, + "grad_norm": 0.019025536254048347, + "learning_rate": 0.0006, + "loss": 4.314227104187012, + "step": 5487 + }, + { + "epoch": 76.2235807860262, + "grad_norm": 0.020918630063533783, + "learning_rate": 0.0006, + "loss": 4.454369068145752, + "step": 5488 + }, + { + "epoch": 76.23755458515284, + "grad_norm": 0.019307617098093033, + "learning_rate": 0.0006, + "loss": 4.247509002685547, + "step": 5489 + }, + { + "epoch": 76.25152838427948, + "grad_norm": 0.014956331811845303, + "learning_rate": 0.0006, + "loss": 4.365175247192383, + "step": 5490 + }, + { + "epoch": 76.26550218340611, + "grad_norm": 0.017213059589266777, + "learning_rate": 0.0006, + "loss": 4.384005069732666, + "step": 5491 + }, + { + "epoch": 76.27947598253274, + "grad_norm": 0.017575597390532494, + "learning_rate": 0.0006, + "loss": 4.41934871673584, + "step": 5492 + }, + { + "epoch": 76.29344978165939, + "grad_norm": 0.015879256650805473, + "learning_rate": 0.0006, + "loss": 4.26675271987915, + "step": 5493 + }, + { + "epoch": 76.30742358078602, + "grad_norm": 0.01564522460103035, + "learning_rate": 0.0006, + "loss": 4.277339935302734, + "step": 5494 + }, + { + "epoch": 76.32139737991267, + "grad_norm": 0.016414586454629898, + "learning_rate": 0.0006, + "loss": 4.4138593673706055, + "step": 5495 + }, + { + "epoch": 76.3353711790393, + "grad_norm": 0.014732337556779385, + "learning_rate": 0.0006, + "loss": 4.281006336212158, + "step": 5496 + }, + { + "epoch": 76.34934497816595, + "grad_norm": 0.015461455099284649, + "learning_rate": 0.0006, + "loss": 4.263422966003418, + "step": 5497 + }, + { + "epoch": 76.36331877729258, + "grad_norm": 0.014529814012348652, + "learning_rate": 0.0006, + "loss": 4.36165189743042, + "step": 5498 + }, + { + "epoch": 76.37729257641921, + "grad_norm": 0.01469483319669962, + "learning_rate": 0.0006, + "loss": 4.309642791748047, + "step": 5499 + }, + { + "epoch": 76.39126637554585, + "grad_norm": 0.016490574926137924, + "learning_rate": 0.0006, + "loss": 4.288150787353516, + "step": 5500 + }, + { + "epoch": 76.40524017467249, + "grad_norm": 0.014593565836548805, + "learning_rate": 0.0006, + "loss": 4.271614074707031, + "step": 5501 + }, + { + "epoch": 76.41921397379913, + "grad_norm": 0.012796535156667233, + "learning_rate": 0.0006, + "loss": 4.420042037963867, + "step": 5502 + }, + { + "epoch": 76.43318777292576, + "grad_norm": 0.016086025163531303, + "learning_rate": 0.0006, + "loss": 4.289943695068359, + "step": 5503 + }, + { + "epoch": 76.44716157205241, + "grad_norm": 0.015439470298588276, + "learning_rate": 0.0006, + "loss": 4.3023271560668945, + "step": 5504 + }, + { + "epoch": 76.46113537117904, + "grad_norm": 0.015066367574036121, + "learning_rate": 0.0006, + "loss": 4.197081565856934, + "step": 5505 + }, + { + "epoch": 76.47510917030567, + "grad_norm": 0.016653263941407204, + "learning_rate": 0.0006, + "loss": 4.333785057067871, + "step": 5506 + }, + { + "epoch": 76.48908296943232, + "grad_norm": 0.014791973866522312, + "learning_rate": 0.0006, + "loss": 4.323811054229736, + "step": 5507 + }, + { + "epoch": 76.50305676855895, + "grad_norm": 0.014694827608764172, + "learning_rate": 0.0006, + "loss": 4.195852279663086, + "step": 5508 + }, + { + "epoch": 76.5170305676856, + "grad_norm": 0.015120396390557289, + "learning_rate": 0.0006, + "loss": 4.42379903793335, + "step": 5509 + }, + { + "epoch": 76.53100436681223, + "grad_norm": 0.014584025368094444, + "learning_rate": 0.0006, + "loss": 4.2702531814575195, + "step": 5510 + }, + { + "epoch": 76.54497816593886, + "grad_norm": 0.013981613330543041, + "learning_rate": 0.0006, + "loss": 4.383029937744141, + "step": 5511 + }, + { + "epoch": 76.5589519650655, + "grad_norm": 0.012930642813444138, + "learning_rate": 0.0006, + "loss": 4.296136856079102, + "step": 5512 + }, + { + "epoch": 76.57292576419214, + "grad_norm": 0.013060989789664745, + "learning_rate": 0.0006, + "loss": 4.346380233764648, + "step": 5513 + }, + { + "epoch": 76.58689956331878, + "grad_norm": 0.013209798373281956, + "learning_rate": 0.0006, + "loss": 4.316067218780518, + "step": 5514 + }, + { + "epoch": 76.60087336244541, + "grad_norm": 0.012423216365277767, + "learning_rate": 0.0006, + "loss": 4.370274543762207, + "step": 5515 + }, + { + "epoch": 76.61484716157206, + "grad_norm": 0.013205411843955517, + "learning_rate": 0.0006, + "loss": 4.429969310760498, + "step": 5516 + }, + { + "epoch": 76.62882096069869, + "grad_norm": 0.013049827888607979, + "learning_rate": 0.0006, + "loss": 4.378549098968506, + "step": 5517 + }, + { + "epoch": 76.64279475982532, + "grad_norm": 0.012075880542397499, + "learning_rate": 0.0006, + "loss": 4.348293304443359, + "step": 5518 + }, + { + "epoch": 76.65676855895197, + "grad_norm": 0.014932847581803799, + "learning_rate": 0.0006, + "loss": 4.282538890838623, + "step": 5519 + }, + { + "epoch": 76.6707423580786, + "grad_norm": 0.015615998767316341, + "learning_rate": 0.0006, + "loss": 4.274641036987305, + "step": 5520 + }, + { + "epoch": 76.68471615720524, + "grad_norm": 0.01424513477832079, + "learning_rate": 0.0006, + "loss": 4.243166923522949, + "step": 5521 + }, + { + "epoch": 76.69868995633188, + "grad_norm": 0.012775546871125698, + "learning_rate": 0.0006, + "loss": 4.335611820220947, + "step": 5522 + }, + { + "epoch": 76.71266375545852, + "grad_norm": 0.014461303129792213, + "learning_rate": 0.0006, + "loss": 4.246976375579834, + "step": 5523 + }, + { + "epoch": 76.72663755458515, + "grad_norm": 0.0145436255261302, + "learning_rate": 0.0006, + "loss": 4.275895118713379, + "step": 5524 + }, + { + "epoch": 76.74061135371178, + "grad_norm": 0.014055085368454456, + "learning_rate": 0.0006, + "loss": 4.331108093261719, + "step": 5525 + }, + { + "epoch": 76.75458515283843, + "grad_norm": 0.011933894827961922, + "learning_rate": 0.0006, + "loss": 4.293721675872803, + "step": 5526 + }, + { + "epoch": 76.76855895196506, + "grad_norm": 0.01277392078191042, + "learning_rate": 0.0006, + "loss": 4.312374114990234, + "step": 5527 + }, + { + "epoch": 76.78253275109171, + "grad_norm": 0.014227936044335365, + "learning_rate": 0.0006, + "loss": 4.238729000091553, + "step": 5528 + }, + { + "epoch": 76.79650655021834, + "grad_norm": 0.01488024927675724, + "learning_rate": 0.0006, + "loss": 4.24587345123291, + "step": 5529 + }, + { + "epoch": 76.81048034934499, + "grad_norm": 0.015341250225901604, + "learning_rate": 0.0006, + "loss": 4.257378578186035, + "step": 5530 + }, + { + "epoch": 76.82445414847162, + "grad_norm": 0.015179437585175037, + "learning_rate": 0.0006, + "loss": 4.2812910079956055, + "step": 5531 + }, + { + "epoch": 76.83842794759825, + "grad_norm": 0.012814724817872047, + "learning_rate": 0.0006, + "loss": 4.2309465408325195, + "step": 5532 + }, + { + "epoch": 76.8524017467249, + "grad_norm": 0.013810716569423676, + "learning_rate": 0.0006, + "loss": 4.325292110443115, + "step": 5533 + }, + { + "epoch": 76.86637554585153, + "grad_norm": 0.017468232661485672, + "learning_rate": 0.0006, + "loss": 4.243788719177246, + "step": 5534 + }, + { + "epoch": 76.88034934497817, + "grad_norm": 0.01739857904613018, + "learning_rate": 0.0006, + "loss": 4.345390796661377, + "step": 5535 + }, + { + "epoch": 76.8943231441048, + "grad_norm": 0.015329061076045036, + "learning_rate": 0.0006, + "loss": 4.415142059326172, + "step": 5536 + }, + { + "epoch": 76.90829694323143, + "grad_norm": 0.01598426327109337, + "learning_rate": 0.0006, + "loss": 4.299558639526367, + "step": 5537 + }, + { + "epoch": 76.92227074235808, + "grad_norm": 0.01621454581618309, + "learning_rate": 0.0006, + "loss": 4.284354209899902, + "step": 5538 + }, + { + "epoch": 76.93624454148471, + "grad_norm": 0.015956968069076538, + "learning_rate": 0.0006, + "loss": 4.293837547302246, + "step": 5539 + }, + { + "epoch": 76.95021834061136, + "grad_norm": 0.01776493526995182, + "learning_rate": 0.0006, + "loss": 4.229729175567627, + "step": 5540 + }, + { + "epoch": 76.96419213973799, + "grad_norm": 0.015317477285861969, + "learning_rate": 0.0006, + "loss": 4.3328447341918945, + "step": 5541 + }, + { + "epoch": 76.97816593886463, + "grad_norm": 0.015140251256525517, + "learning_rate": 0.0006, + "loss": 4.310840606689453, + "step": 5542 + }, + { + "epoch": 76.99213973799127, + "grad_norm": 0.015326812863349915, + "learning_rate": 0.0006, + "loss": 4.230832099914551, + "step": 5543 + }, + { + "epoch": 77.0, + "grad_norm": 0.01607763208448887, + "learning_rate": 0.0006, + "loss": 4.2236328125, + "step": 5544 + }, + { + "epoch": 77.0, + "eval_loss": 4.662221431732178, + "eval_runtime": 56.5194, + "eval_samples_per_second": 43.206, + "eval_steps_per_second": 1.362, + "step": 5544 + }, + { + "epoch": 77.01397379912663, + "grad_norm": 0.014839510433375835, + "learning_rate": 0.0006, + "loss": 4.254111289978027, + "step": 5545 + }, + { + "epoch": 77.02794759825328, + "grad_norm": 0.01519741676747799, + "learning_rate": 0.0006, + "loss": 4.1607184410095215, + "step": 5546 + }, + { + "epoch": 77.04192139737991, + "grad_norm": 0.015472603030502796, + "learning_rate": 0.0006, + "loss": 4.323320388793945, + "step": 5547 + }, + { + "epoch": 77.05589519650655, + "grad_norm": 0.017937948927283287, + "learning_rate": 0.0006, + "loss": 4.308490753173828, + "step": 5548 + }, + { + "epoch": 77.06986899563319, + "grad_norm": 0.020890841260552406, + "learning_rate": 0.0006, + "loss": 4.259101867675781, + "step": 5549 + }, + { + "epoch": 77.08384279475983, + "grad_norm": 0.01892857998609543, + "learning_rate": 0.0006, + "loss": 4.386970520019531, + "step": 5550 + }, + { + "epoch": 77.09781659388646, + "grad_norm": 0.016583630815148354, + "learning_rate": 0.0006, + "loss": 4.233604907989502, + "step": 5551 + }, + { + "epoch": 77.1117903930131, + "grad_norm": 0.016063468530774117, + "learning_rate": 0.0006, + "loss": 4.339544296264648, + "step": 5552 + }, + { + "epoch": 77.12576419213974, + "grad_norm": 0.018326779827475548, + "learning_rate": 0.0006, + "loss": 4.367204666137695, + "step": 5553 + }, + { + "epoch": 77.13973799126637, + "grad_norm": 0.021590933203697205, + "learning_rate": 0.0006, + "loss": 4.352685928344727, + "step": 5554 + }, + { + "epoch": 77.15371179039302, + "grad_norm": 0.019643642008304596, + "learning_rate": 0.0006, + "loss": 4.258535385131836, + "step": 5555 + }, + { + "epoch": 77.16768558951965, + "grad_norm": 0.018476149067282677, + "learning_rate": 0.0006, + "loss": 4.232568740844727, + "step": 5556 + }, + { + "epoch": 77.18165938864628, + "grad_norm": 0.01584405079483986, + "learning_rate": 0.0006, + "loss": 4.258371353149414, + "step": 5557 + }, + { + "epoch": 77.19563318777293, + "grad_norm": 0.01575201191008091, + "learning_rate": 0.0006, + "loss": 4.183218002319336, + "step": 5558 + }, + { + "epoch": 77.20960698689956, + "grad_norm": 0.016871824860572815, + "learning_rate": 0.0006, + "loss": 4.194563865661621, + "step": 5559 + }, + { + "epoch": 77.2235807860262, + "grad_norm": 0.02059333398938179, + "learning_rate": 0.0006, + "loss": 4.195281982421875, + "step": 5560 + }, + { + "epoch": 77.23755458515284, + "grad_norm": 0.022640453651547432, + "learning_rate": 0.0006, + "loss": 4.279615879058838, + "step": 5561 + }, + { + "epoch": 77.25152838427948, + "grad_norm": 0.020335959270596504, + "learning_rate": 0.0006, + "loss": 4.25040864944458, + "step": 5562 + }, + { + "epoch": 77.26550218340611, + "grad_norm": 0.020010793581604958, + "learning_rate": 0.0006, + "loss": 4.291116714477539, + "step": 5563 + }, + { + "epoch": 77.27947598253274, + "grad_norm": 0.02109045907855034, + "learning_rate": 0.0006, + "loss": 4.364088535308838, + "step": 5564 + }, + { + "epoch": 77.29344978165939, + "grad_norm": 0.019519446417689323, + "learning_rate": 0.0006, + "loss": 4.329166412353516, + "step": 5565 + }, + { + "epoch": 77.30742358078602, + "grad_norm": 0.01689804159104824, + "learning_rate": 0.0006, + "loss": 4.244176864624023, + "step": 5566 + }, + { + "epoch": 77.32139737991267, + "grad_norm": 0.01893681474030018, + "learning_rate": 0.0006, + "loss": 4.239068984985352, + "step": 5567 + }, + { + "epoch": 77.3353711790393, + "grad_norm": 0.015998193994164467, + "learning_rate": 0.0006, + "loss": 4.173398017883301, + "step": 5568 + }, + { + "epoch": 77.34934497816595, + "grad_norm": 0.013892280869185925, + "learning_rate": 0.0006, + "loss": 4.287654876708984, + "step": 5569 + }, + { + "epoch": 77.36331877729258, + "grad_norm": 0.017060643061995506, + "learning_rate": 0.0006, + "loss": 4.398447036743164, + "step": 5570 + }, + { + "epoch": 77.37729257641921, + "grad_norm": 0.015168849378824234, + "learning_rate": 0.0006, + "loss": 4.15665864944458, + "step": 5571 + }, + { + "epoch": 77.39126637554585, + "grad_norm": 0.014886523596942425, + "learning_rate": 0.0006, + "loss": 4.196673393249512, + "step": 5572 + }, + { + "epoch": 77.40524017467249, + "grad_norm": 0.016941796988248825, + "learning_rate": 0.0006, + "loss": 4.314562797546387, + "step": 5573 + }, + { + "epoch": 77.41921397379913, + "grad_norm": 0.015353812836110592, + "learning_rate": 0.0006, + "loss": 4.409579277038574, + "step": 5574 + }, + { + "epoch": 77.43318777292576, + "grad_norm": 0.015192852355539799, + "learning_rate": 0.0006, + "loss": 4.392045974731445, + "step": 5575 + }, + { + "epoch": 77.44716157205241, + "grad_norm": 0.014232569374144077, + "learning_rate": 0.0006, + "loss": 4.21639347076416, + "step": 5576 + }, + { + "epoch": 77.46113537117904, + "grad_norm": 0.014453226700425148, + "learning_rate": 0.0006, + "loss": 4.2802276611328125, + "step": 5577 + }, + { + "epoch": 77.47510917030567, + "grad_norm": 0.012881236150860786, + "learning_rate": 0.0006, + "loss": 4.375980854034424, + "step": 5578 + }, + { + "epoch": 77.48908296943232, + "grad_norm": 0.01583976298570633, + "learning_rate": 0.0006, + "loss": 4.264520645141602, + "step": 5579 + }, + { + "epoch": 77.50305676855895, + "grad_norm": 0.014171584509313107, + "learning_rate": 0.0006, + "loss": 4.225279808044434, + "step": 5580 + }, + { + "epoch": 77.5170305676856, + "grad_norm": 0.014715159311890602, + "learning_rate": 0.0006, + "loss": 4.369834899902344, + "step": 5581 + }, + { + "epoch": 77.53100436681223, + "grad_norm": 0.016554079949855804, + "learning_rate": 0.0006, + "loss": 4.2976179122924805, + "step": 5582 + }, + { + "epoch": 77.54497816593886, + "grad_norm": 0.015095747075974941, + "learning_rate": 0.0006, + "loss": 4.294407844543457, + "step": 5583 + }, + { + "epoch": 77.5589519650655, + "grad_norm": 0.012726361863315105, + "learning_rate": 0.0006, + "loss": 4.221401214599609, + "step": 5584 + }, + { + "epoch": 77.57292576419214, + "grad_norm": 0.014999309554696083, + "learning_rate": 0.0006, + "loss": 4.243929386138916, + "step": 5585 + }, + { + "epoch": 77.58689956331878, + "grad_norm": 0.016526037827134132, + "learning_rate": 0.0006, + "loss": 4.374139785766602, + "step": 5586 + }, + { + "epoch": 77.60087336244541, + "grad_norm": 0.01648339442908764, + "learning_rate": 0.0006, + "loss": 4.392425060272217, + "step": 5587 + }, + { + "epoch": 77.61484716157206, + "grad_norm": 0.015269213356077671, + "learning_rate": 0.0006, + "loss": 4.33986759185791, + "step": 5588 + }, + { + "epoch": 77.62882096069869, + "grad_norm": 0.01690507121384144, + "learning_rate": 0.0006, + "loss": 4.307476043701172, + "step": 5589 + }, + { + "epoch": 77.64279475982532, + "grad_norm": 0.016438685357570648, + "learning_rate": 0.0006, + "loss": 4.2406206130981445, + "step": 5590 + }, + { + "epoch": 77.65676855895197, + "grad_norm": 0.0157600287348032, + "learning_rate": 0.0006, + "loss": 4.314070224761963, + "step": 5591 + }, + { + "epoch": 77.6707423580786, + "grad_norm": 0.015127947553992271, + "learning_rate": 0.0006, + "loss": 4.22475528717041, + "step": 5592 + }, + { + "epoch": 77.68471615720524, + "grad_norm": 0.016632677987217903, + "learning_rate": 0.0006, + "loss": 4.26954984664917, + "step": 5593 + }, + { + "epoch": 77.69868995633188, + "grad_norm": 0.013150460086762905, + "learning_rate": 0.0006, + "loss": 4.369131088256836, + "step": 5594 + }, + { + "epoch": 77.71266375545852, + "grad_norm": 0.014513005502521992, + "learning_rate": 0.0006, + "loss": 4.3248796463012695, + "step": 5595 + }, + { + "epoch": 77.72663755458515, + "grad_norm": 0.014090991578996181, + "learning_rate": 0.0006, + "loss": 4.268381118774414, + "step": 5596 + }, + { + "epoch": 77.74061135371178, + "grad_norm": 0.01633097417652607, + "learning_rate": 0.0006, + "loss": 4.238439083099365, + "step": 5597 + }, + { + "epoch": 77.75458515283843, + "grad_norm": 0.01710933819413185, + "learning_rate": 0.0006, + "loss": 4.399697780609131, + "step": 5598 + }, + { + "epoch": 77.76855895196506, + "grad_norm": 0.019251767545938492, + "learning_rate": 0.0006, + "loss": 4.290450096130371, + "step": 5599 + }, + { + "epoch": 77.78253275109171, + "grad_norm": 0.01849171705543995, + "learning_rate": 0.0006, + "loss": 4.267734527587891, + "step": 5600 + }, + { + "epoch": 77.79650655021834, + "grad_norm": 0.014173115603625774, + "learning_rate": 0.0006, + "loss": 4.288799285888672, + "step": 5601 + }, + { + "epoch": 77.81048034934499, + "grad_norm": 0.014572718180716038, + "learning_rate": 0.0006, + "loss": 4.3790435791015625, + "step": 5602 + }, + { + "epoch": 77.82445414847162, + "grad_norm": 0.016375336796045303, + "learning_rate": 0.0006, + "loss": 4.231725692749023, + "step": 5603 + }, + { + "epoch": 77.83842794759825, + "grad_norm": 0.017326701432466507, + "learning_rate": 0.0006, + "loss": 4.211495399475098, + "step": 5604 + }, + { + "epoch": 77.8524017467249, + "grad_norm": 0.016399390995502472, + "learning_rate": 0.0006, + "loss": 4.386699199676514, + "step": 5605 + }, + { + "epoch": 77.86637554585153, + "grad_norm": 0.018269989639520645, + "learning_rate": 0.0006, + "loss": 4.254856109619141, + "step": 5606 + }, + { + "epoch": 77.88034934497817, + "grad_norm": 0.018443435430526733, + "learning_rate": 0.0006, + "loss": 4.391961097717285, + "step": 5607 + }, + { + "epoch": 77.8943231441048, + "grad_norm": 0.01784636080265045, + "learning_rate": 0.0006, + "loss": 4.273881912231445, + "step": 5608 + }, + { + "epoch": 77.90829694323143, + "grad_norm": 0.01782994158565998, + "learning_rate": 0.0006, + "loss": 4.26833963394165, + "step": 5609 + }, + { + "epoch": 77.92227074235808, + "grad_norm": 0.016260741278529167, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 5610 + }, + { + "epoch": 77.93624454148471, + "grad_norm": 0.015482903458178043, + "learning_rate": 0.0006, + "loss": 4.187826633453369, + "step": 5611 + }, + { + "epoch": 77.95021834061136, + "grad_norm": 0.015193257480859756, + "learning_rate": 0.0006, + "loss": 4.286060333251953, + "step": 5612 + }, + { + "epoch": 77.96419213973799, + "grad_norm": 0.015110115520656109, + "learning_rate": 0.0006, + "loss": 4.2828192710876465, + "step": 5613 + }, + { + "epoch": 77.97816593886463, + "grad_norm": 0.015973486006259918, + "learning_rate": 0.0006, + "loss": 4.317939758300781, + "step": 5614 + }, + { + "epoch": 77.99213973799127, + "grad_norm": 0.017363594844937325, + "learning_rate": 0.0006, + "loss": 4.287721633911133, + "step": 5615 + }, + { + "epoch": 78.0, + "grad_norm": 0.015277642756700516, + "learning_rate": 0.0006, + "loss": 4.356618881225586, + "step": 5616 + }, + { + "epoch": 78.0, + "eval_loss": 4.642678737640381, + "eval_runtime": 56.7335, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 5616 + }, + { + "epoch": 78.01397379912663, + "grad_norm": 0.016794051975011826, + "learning_rate": 0.0006, + "loss": 4.279930591583252, + "step": 5617 + }, + { + "epoch": 78.02794759825328, + "grad_norm": 0.01632673107087612, + "learning_rate": 0.0006, + "loss": 4.235943794250488, + "step": 5618 + }, + { + "epoch": 78.04192139737991, + "grad_norm": 0.01437876746058464, + "learning_rate": 0.0006, + "loss": 4.2542243003845215, + "step": 5619 + }, + { + "epoch": 78.05589519650655, + "grad_norm": 0.014993119053542614, + "learning_rate": 0.0006, + "loss": 4.262046813964844, + "step": 5620 + }, + { + "epoch": 78.06986899563319, + "grad_norm": 0.015721119940280914, + "learning_rate": 0.0006, + "loss": 4.173436164855957, + "step": 5621 + }, + { + "epoch": 78.08384279475983, + "grad_norm": 0.01334085687994957, + "learning_rate": 0.0006, + "loss": 4.322175979614258, + "step": 5622 + }, + { + "epoch": 78.09781659388646, + "grad_norm": 0.015561388805508614, + "learning_rate": 0.0006, + "loss": 4.242125988006592, + "step": 5623 + }, + { + "epoch": 78.1117903930131, + "grad_norm": 0.014418602921068668, + "learning_rate": 0.0006, + "loss": 4.266186714172363, + "step": 5624 + }, + { + "epoch": 78.12576419213974, + "grad_norm": 0.013739909045398235, + "learning_rate": 0.0006, + "loss": 4.231633186340332, + "step": 5625 + }, + { + "epoch": 78.13973799126637, + "grad_norm": 0.01402067206799984, + "learning_rate": 0.0006, + "loss": 4.188406944274902, + "step": 5626 + }, + { + "epoch": 78.15371179039302, + "grad_norm": 0.013639481738209724, + "learning_rate": 0.0006, + "loss": 4.1586594581604, + "step": 5627 + }, + { + "epoch": 78.16768558951965, + "grad_norm": 0.012858159840106964, + "learning_rate": 0.0006, + "loss": 4.360448360443115, + "step": 5628 + }, + { + "epoch": 78.18165938864628, + "grad_norm": 0.013089739717543125, + "learning_rate": 0.0006, + "loss": 4.344371795654297, + "step": 5629 + }, + { + "epoch": 78.19563318777293, + "grad_norm": 0.01382314320653677, + "learning_rate": 0.0006, + "loss": 4.20378303527832, + "step": 5630 + }, + { + "epoch": 78.20960698689956, + "grad_norm": 0.014206619001924992, + "learning_rate": 0.0006, + "loss": 4.239891529083252, + "step": 5631 + }, + { + "epoch": 78.2235807860262, + "grad_norm": 0.01472168043255806, + "learning_rate": 0.0006, + "loss": 4.282382488250732, + "step": 5632 + }, + { + "epoch": 78.23755458515284, + "grad_norm": 0.014237025752663612, + "learning_rate": 0.0006, + "loss": 4.241545677185059, + "step": 5633 + }, + { + "epoch": 78.25152838427948, + "grad_norm": 0.013231472112238407, + "learning_rate": 0.0006, + "loss": 4.184896469116211, + "step": 5634 + }, + { + "epoch": 78.26550218340611, + "grad_norm": 0.013538471423089504, + "learning_rate": 0.0006, + "loss": 4.325181007385254, + "step": 5635 + }, + { + "epoch": 78.27947598253274, + "grad_norm": 0.013541961088776588, + "learning_rate": 0.0006, + "loss": 4.368722915649414, + "step": 5636 + }, + { + "epoch": 78.29344978165939, + "grad_norm": 0.013446723110973835, + "learning_rate": 0.0006, + "loss": 4.301422595977783, + "step": 5637 + }, + { + "epoch": 78.30742358078602, + "grad_norm": 0.014753809198737144, + "learning_rate": 0.0006, + "loss": 4.259127616882324, + "step": 5638 + }, + { + "epoch": 78.32139737991267, + "grad_norm": 0.013256008736789227, + "learning_rate": 0.0006, + "loss": 4.284032821655273, + "step": 5639 + }, + { + "epoch": 78.3353711790393, + "grad_norm": 0.014272629283368587, + "learning_rate": 0.0006, + "loss": 4.234989166259766, + "step": 5640 + }, + { + "epoch": 78.34934497816595, + "grad_norm": 0.014756826683878899, + "learning_rate": 0.0006, + "loss": 4.143499374389648, + "step": 5641 + }, + { + "epoch": 78.36331877729258, + "grad_norm": 0.016549205407500267, + "learning_rate": 0.0006, + "loss": 4.311034202575684, + "step": 5642 + }, + { + "epoch": 78.37729257641921, + "grad_norm": 0.017735283821821213, + "learning_rate": 0.0006, + "loss": 4.246767520904541, + "step": 5643 + }, + { + "epoch": 78.39126637554585, + "grad_norm": 0.017370613291859627, + "learning_rate": 0.0006, + "loss": 4.2460832595825195, + "step": 5644 + }, + { + "epoch": 78.40524017467249, + "grad_norm": 0.014642846770584583, + "learning_rate": 0.0006, + "loss": 4.303811550140381, + "step": 5645 + }, + { + "epoch": 78.41921397379913, + "grad_norm": 0.013445726595818996, + "learning_rate": 0.0006, + "loss": 4.25921630859375, + "step": 5646 + }, + { + "epoch": 78.43318777292576, + "grad_norm": 0.014254065230488777, + "learning_rate": 0.0006, + "loss": 4.129446506500244, + "step": 5647 + }, + { + "epoch": 78.44716157205241, + "grad_norm": 0.01883019506931305, + "learning_rate": 0.0006, + "loss": 4.191716194152832, + "step": 5648 + }, + { + "epoch": 78.46113537117904, + "grad_norm": 0.0204437468200922, + "learning_rate": 0.0006, + "loss": 4.333085536956787, + "step": 5649 + }, + { + "epoch": 78.47510917030567, + "grad_norm": 0.01652318798005581, + "learning_rate": 0.0006, + "loss": 4.135462760925293, + "step": 5650 + }, + { + "epoch": 78.48908296943232, + "grad_norm": 0.012753551825881004, + "learning_rate": 0.0006, + "loss": 4.195051193237305, + "step": 5651 + }, + { + "epoch": 78.50305676855895, + "grad_norm": 0.01499118935316801, + "learning_rate": 0.0006, + "loss": 4.157092094421387, + "step": 5652 + }, + { + "epoch": 78.5170305676856, + "grad_norm": 0.015383945778012276, + "learning_rate": 0.0006, + "loss": 4.253818035125732, + "step": 5653 + }, + { + "epoch": 78.53100436681223, + "grad_norm": 0.015721332281827927, + "learning_rate": 0.0006, + "loss": 4.228775978088379, + "step": 5654 + }, + { + "epoch": 78.54497816593886, + "grad_norm": 0.014595463871955872, + "learning_rate": 0.0006, + "loss": 4.135197639465332, + "step": 5655 + }, + { + "epoch": 78.5589519650655, + "grad_norm": 0.015059644356369972, + "learning_rate": 0.0006, + "loss": 4.255614280700684, + "step": 5656 + }, + { + "epoch": 78.57292576419214, + "grad_norm": 0.013833344914019108, + "learning_rate": 0.0006, + "loss": 4.301250457763672, + "step": 5657 + }, + { + "epoch": 78.58689956331878, + "grad_norm": 0.013665376231074333, + "learning_rate": 0.0006, + "loss": 4.1734232902526855, + "step": 5658 + }, + { + "epoch": 78.60087336244541, + "grad_norm": 0.01539086177945137, + "learning_rate": 0.0006, + "loss": 4.221459865570068, + "step": 5659 + }, + { + "epoch": 78.61484716157206, + "grad_norm": 0.013313917443156242, + "learning_rate": 0.0006, + "loss": 4.255619049072266, + "step": 5660 + }, + { + "epoch": 78.62882096069869, + "grad_norm": 0.01447741687297821, + "learning_rate": 0.0006, + "loss": 4.298219680786133, + "step": 5661 + }, + { + "epoch": 78.64279475982532, + "grad_norm": 0.012796915136277676, + "learning_rate": 0.0006, + "loss": 4.2542405128479, + "step": 5662 + }, + { + "epoch": 78.65676855895197, + "grad_norm": 0.014058141969144344, + "learning_rate": 0.0006, + "loss": 4.284477233886719, + "step": 5663 + }, + { + "epoch": 78.6707423580786, + "grad_norm": 0.012705069966614246, + "learning_rate": 0.0006, + "loss": 4.2826666831970215, + "step": 5664 + }, + { + "epoch": 78.68471615720524, + "grad_norm": 0.013245890848338604, + "learning_rate": 0.0006, + "loss": 4.371612548828125, + "step": 5665 + }, + { + "epoch": 78.69868995633188, + "grad_norm": 0.014192677102982998, + "learning_rate": 0.0006, + "loss": 4.267344951629639, + "step": 5666 + }, + { + "epoch": 78.71266375545852, + "grad_norm": 0.015346777625381947, + "learning_rate": 0.0006, + "loss": 4.222797393798828, + "step": 5667 + }, + { + "epoch": 78.72663755458515, + "grad_norm": 0.014068621210753918, + "learning_rate": 0.0006, + "loss": 4.172964572906494, + "step": 5668 + }, + { + "epoch": 78.74061135371178, + "grad_norm": 0.014364945702254772, + "learning_rate": 0.0006, + "loss": 4.259319305419922, + "step": 5669 + }, + { + "epoch": 78.75458515283843, + "grad_norm": 0.017337387427687645, + "learning_rate": 0.0006, + "loss": 4.29533576965332, + "step": 5670 + }, + { + "epoch": 78.76855895196506, + "grad_norm": 0.017416715621948242, + "learning_rate": 0.0006, + "loss": 4.340364456176758, + "step": 5671 + }, + { + "epoch": 78.78253275109171, + "grad_norm": 0.015925846993923187, + "learning_rate": 0.0006, + "loss": 4.177454948425293, + "step": 5672 + }, + { + "epoch": 78.79650655021834, + "grad_norm": 0.013130443170666695, + "learning_rate": 0.0006, + "loss": 4.277454853057861, + "step": 5673 + }, + { + "epoch": 78.81048034934499, + "grad_norm": 0.014628687873482704, + "learning_rate": 0.0006, + "loss": 4.273757457733154, + "step": 5674 + }, + { + "epoch": 78.82445414847162, + "grad_norm": 0.01601659320294857, + "learning_rate": 0.0006, + "loss": 4.281774520874023, + "step": 5675 + }, + { + "epoch": 78.83842794759825, + "grad_norm": 0.015443485230207443, + "learning_rate": 0.0006, + "loss": 4.199937343597412, + "step": 5676 + }, + { + "epoch": 78.8524017467249, + "grad_norm": 0.013934246264398098, + "learning_rate": 0.0006, + "loss": 4.331369876861572, + "step": 5677 + }, + { + "epoch": 78.86637554585153, + "grad_norm": 0.01572083681821823, + "learning_rate": 0.0006, + "loss": 4.2828474044799805, + "step": 5678 + }, + { + "epoch": 78.88034934497817, + "grad_norm": 0.013444120064377785, + "learning_rate": 0.0006, + "loss": 4.183859825134277, + "step": 5679 + }, + { + "epoch": 78.8943231441048, + "grad_norm": 0.01591254211962223, + "learning_rate": 0.0006, + "loss": 4.295256614685059, + "step": 5680 + }, + { + "epoch": 78.90829694323143, + "grad_norm": 0.017556684091687202, + "learning_rate": 0.0006, + "loss": 4.132272720336914, + "step": 5681 + }, + { + "epoch": 78.92227074235808, + "grad_norm": 0.018184322863817215, + "learning_rate": 0.0006, + "loss": 4.251728057861328, + "step": 5682 + }, + { + "epoch": 78.93624454148471, + "grad_norm": 0.016207285225391388, + "learning_rate": 0.0006, + "loss": 4.205544471740723, + "step": 5683 + }, + { + "epoch": 78.95021834061136, + "grad_norm": 0.014478376135230064, + "learning_rate": 0.0006, + "loss": 4.182462215423584, + "step": 5684 + }, + { + "epoch": 78.96419213973799, + "grad_norm": 0.014255235902965069, + "learning_rate": 0.0006, + "loss": 4.253366947174072, + "step": 5685 + }, + { + "epoch": 78.97816593886463, + "grad_norm": 0.01524396613240242, + "learning_rate": 0.0006, + "loss": 4.221786022186279, + "step": 5686 + }, + { + "epoch": 78.99213973799127, + "grad_norm": 0.015057693235576153, + "learning_rate": 0.0006, + "loss": 4.208454608917236, + "step": 5687 + }, + { + "epoch": 79.0, + "grad_norm": 0.015478022396564484, + "learning_rate": 0.0006, + "loss": 4.336383819580078, + "step": 5688 + }, + { + "epoch": 79.0, + "eval_loss": 4.641749382019043, + "eval_runtime": 56.8167, + "eval_samples_per_second": 42.98, + "eval_steps_per_second": 1.355, + "step": 5688 + }, + { + "epoch": 79.01397379912663, + "grad_norm": 0.014143316075205803, + "learning_rate": 0.0006, + "loss": 4.260974884033203, + "step": 5689 + }, + { + "epoch": 79.02794759825328, + "grad_norm": 0.013120883144438267, + "learning_rate": 0.0006, + "loss": 4.139886856079102, + "step": 5690 + }, + { + "epoch": 79.04192139737991, + "grad_norm": 0.015019661746919155, + "learning_rate": 0.0006, + "loss": 4.267004013061523, + "step": 5691 + }, + { + "epoch": 79.05589519650655, + "grad_norm": 0.014891284517943859, + "learning_rate": 0.0006, + "loss": 4.205019950866699, + "step": 5692 + }, + { + "epoch": 79.06986899563319, + "grad_norm": 0.015527728945016861, + "learning_rate": 0.0006, + "loss": 4.1513566970825195, + "step": 5693 + }, + { + "epoch": 79.08384279475983, + "grad_norm": 0.017527181655168533, + "learning_rate": 0.0006, + "loss": 4.148828983306885, + "step": 5694 + }, + { + "epoch": 79.09781659388646, + "grad_norm": 0.01632206328213215, + "learning_rate": 0.0006, + "loss": 4.195789813995361, + "step": 5695 + }, + { + "epoch": 79.1117903930131, + "grad_norm": 0.01590503193438053, + "learning_rate": 0.0006, + "loss": 4.199155330657959, + "step": 5696 + }, + { + "epoch": 79.12576419213974, + "grad_norm": 0.016489997506141663, + "learning_rate": 0.0006, + "loss": 4.155154705047607, + "step": 5697 + }, + { + "epoch": 79.13973799126637, + "grad_norm": 0.0161368977278471, + "learning_rate": 0.0006, + "loss": 4.164384841918945, + "step": 5698 + }, + { + "epoch": 79.15371179039302, + "grad_norm": 0.018574809655547142, + "learning_rate": 0.0006, + "loss": 4.1898193359375, + "step": 5699 + }, + { + "epoch": 79.16768558951965, + "grad_norm": 0.018987983465194702, + "learning_rate": 0.0006, + "loss": 4.219719409942627, + "step": 5700 + }, + { + "epoch": 79.18165938864628, + "grad_norm": 0.01668606884777546, + "learning_rate": 0.0006, + "loss": 4.152734756469727, + "step": 5701 + }, + { + "epoch": 79.19563318777293, + "grad_norm": 0.014046299271285534, + "learning_rate": 0.0006, + "loss": 4.152964115142822, + "step": 5702 + }, + { + "epoch": 79.20960698689956, + "grad_norm": 0.01628032699227333, + "learning_rate": 0.0006, + "loss": 4.299479961395264, + "step": 5703 + }, + { + "epoch": 79.2235807860262, + "grad_norm": 0.017998795956373215, + "learning_rate": 0.0006, + "loss": 4.231203079223633, + "step": 5704 + }, + { + "epoch": 79.23755458515284, + "grad_norm": 0.014799647964537144, + "learning_rate": 0.0006, + "loss": 4.3530192375183105, + "step": 5705 + }, + { + "epoch": 79.25152838427948, + "grad_norm": 0.014640678651630878, + "learning_rate": 0.0006, + "loss": 4.287322998046875, + "step": 5706 + }, + { + "epoch": 79.26550218340611, + "grad_norm": 0.0172616895288229, + "learning_rate": 0.0006, + "loss": 4.148009777069092, + "step": 5707 + }, + { + "epoch": 79.27947598253274, + "grad_norm": 0.01811790093779564, + "learning_rate": 0.0006, + "loss": 4.269433975219727, + "step": 5708 + }, + { + "epoch": 79.29344978165939, + "grad_norm": 0.01764000579714775, + "learning_rate": 0.0006, + "loss": 4.292638778686523, + "step": 5709 + }, + { + "epoch": 79.30742358078602, + "grad_norm": 0.018030935898423195, + "learning_rate": 0.0006, + "loss": 4.267889976501465, + "step": 5710 + }, + { + "epoch": 79.32139737991267, + "grad_norm": 0.017860891297459602, + "learning_rate": 0.0006, + "loss": 4.211110591888428, + "step": 5711 + }, + { + "epoch": 79.3353711790393, + "grad_norm": 0.017295073717832565, + "learning_rate": 0.0006, + "loss": 4.224645614624023, + "step": 5712 + }, + { + "epoch": 79.34934497816595, + "grad_norm": 0.015891101211309433, + "learning_rate": 0.0006, + "loss": 4.224797248840332, + "step": 5713 + }, + { + "epoch": 79.36331877729258, + "grad_norm": 0.014680417254567146, + "learning_rate": 0.0006, + "loss": 4.182242393493652, + "step": 5714 + }, + { + "epoch": 79.37729257641921, + "grad_norm": 0.016621140763163567, + "learning_rate": 0.0006, + "loss": 4.186334609985352, + "step": 5715 + }, + { + "epoch": 79.39126637554585, + "grad_norm": 0.016207249835133553, + "learning_rate": 0.0006, + "loss": 4.277074813842773, + "step": 5716 + }, + { + "epoch": 79.40524017467249, + "grad_norm": 0.015375478193163872, + "learning_rate": 0.0006, + "loss": 4.154317855834961, + "step": 5717 + }, + { + "epoch": 79.41921397379913, + "grad_norm": 0.013453228399157524, + "learning_rate": 0.0006, + "loss": 4.265270233154297, + "step": 5718 + }, + { + "epoch": 79.43318777292576, + "grad_norm": 0.014312694780528545, + "learning_rate": 0.0006, + "loss": 4.232220649719238, + "step": 5719 + }, + { + "epoch": 79.44716157205241, + "grad_norm": 0.014759554527699947, + "learning_rate": 0.0006, + "loss": 4.272275924682617, + "step": 5720 + }, + { + "epoch": 79.46113537117904, + "grad_norm": 0.015885505825281143, + "learning_rate": 0.0006, + "loss": 4.2229719161987305, + "step": 5721 + }, + { + "epoch": 79.47510917030567, + "grad_norm": 0.016516581177711487, + "learning_rate": 0.0006, + "loss": 4.200624465942383, + "step": 5722 + }, + { + "epoch": 79.48908296943232, + "grad_norm": 0.018254555761814117, + "learning_rate": 0.0006, + "loss": 4.19648551940918, + "step": 5723 + }, + { + "epoch": 79.50305676855895, + "grad_norm": 0.01467090006917715, + "learning_rate": 0.0006, + "loss": 4.135190486907959, + "step": 5724 + }, + { + "epoch": 79.5170305676856, + "grad_norm": 0.013474797829985619, + "learning_rate": 0.0006, + "loss": 4.311361312866211, + "step": 5725 + }, + { + "epoch": 79.53100436681223, + "grad_norm": 0.01742764748632908, + "learning_rate": 0.0006, + "loss": 4.1845903396606445, + "step": 5726 + }, + { + "epoch": 79.54497816593886, + "grad_norm": 0.018786882981657982, + "learning_rate": 0.0006, + "loss": 4.163942337036133, + "step": 5727 + }, + { + "epoch": 79.5589519650655, + "grad_norm": 0.017831219360232353, + "learning_rate": 0.0006, + "loss": 4.236685276031494, + "step": 5728 + }, + { + "epoch": 79.57292576419214, + "grad_norm": 0.016334986314177513, + "learning_rate": 0.0006, + "loss": 4.255853652954102, + "step": 5729 + }, + { + "epoch": 79.58689956331878, + "grad_norm": 0.0138487434014678, + "learning_rate": 0.0006, + "loss": 4.227874279022217, + "step": 5730 + }, + { + "epoch": 79.60087336244541, + "grad_norm": 0.01531760673969984, + "learning_rate": 0.0006, + "loss": 4.221432685852051, + "step": 5731 + }, + { + "epoch": 79.61484716157206, + "grad_norm": 0.01603361964225769, + "learning_rate": 0.0006, + "loss": 4.278326511383057, + "step": 5732 + }, + { + "epoch": 79.62882096069869, + "grad_norm": 0.013149895705282688, + "learning_rate": 0.0006, + "loss": 4.2474517822265625, + "step": 5733 + }, + { + "epoch": 79.64279475982532, + "grad_norm": 0.013428786769509315, + "learning_rate": 0.0006, + "loss": 4.1701178550720215, + "step": 5734 + }, + { + "epoch": 79.65676855895197, + "grad_norm": 0.014305585995316505, + "learning_rate": 0.0006, + "loss": 4.250778675079346, + "step": 5735 + }, + { + "epoch": 79.6707423580786, + "grad_norm": 0.014544566161930561, + "learning_rate": 0.0006, + "loss": 4.243236541748047, + "step": 5736 + }, + { + "epoch": 79.68471615720524, + "grad_norm": 0.013524414040148258, + "learning_rate": 0.0006, + "loss": 4.255457878112793, + "step": 5737 + }, + { + "epoch": 79.69868995633188, + "grad_norm": 0.014750641770660877, + "learning_rate": 0.0006, + "loss": 4.202816486358643, + "step": 5738 + }, + { + "epoch": 79.71266375545852, + "grad_norm": 0.015090773813426495, + "learning_rate": 0.0006, + "loss": 4.3053178787231445, + "step": 5739 + }, + { + "epoch": 79.72663755458515, + "grad_norm": 0.01606169529259205, + "learning_rate": 0.0006, + "loss": 4.224017143249512, + "step": 5740 + }, + { + "epoch": 79.74061135371178, + "grad_norm": 0.016303520649671555, + "learning_rate": 0.0006, + "loss": 4.190760612487793, + "step": 5741 + }, + { + "epoch": 79.75458515283843, + "grad_norm": 0.01465672068297863, + "learning_rate": 0.0006, + "loss": 4.24506950378418, + "step": 5742 + }, + { + "epoch": 79.76855895196506, + "grad_norm": 0.012616301886737347, + "learning_rate": 0.0006, + "loss": 4.174039840698242, + "step": 5743 + }, + { + "epoch": 79.78253275109171, + "grad_norm": 0.01358284242451191, + "learning_rate": 0.0006, + "loss": 4.236380577087402, + "step": 5744 + }, + { + "epoch": 79.79650655021834, + "grad_norm": 0.01557177770882845, + "learning_rate": 0.0006, + "loss": 4.102962493896484, + "step": 5745 + }, + { + "epoch": 79.81048034934499, + "grad_norm": 0.015584317035973072, + "learning_rate": 0.0006, + "loss": 4.242603302001953, + "step": 5746 + }, + { + "epoch": 79.82445414847162, + "grad_norm": 0.012597334571182728, + "learning_rate": 0.0006, + "loss": 4.202791690826416, + "step": 5747 + }, + { + "epoch": 79.83842794759825, + "grad_norm": 0.013722898438572884, + "learning_rate": 0.0006, + "loss": 4.204328536987305, + "step": 5748 + }, + { + "epoch": 79.8524017467249, + "grad_norm": 0.014669225551187992, + "learning_rate": 0.0006, + "loss": 4.164331912994385, + "step": 5749 + }, + { + "epoch": 79.86637554585153, + "grad_norm": 0.015012740157544613, + "learning_rate": 0.0006, + "loss": 4.260565280914307, + "step": 5750 + }, + { + "epoch": 79.88034934497817, + "grad_norm": 0.016026340425014496, + "learning_rate": 0.0006, + "loss": 4.299723148345947, + "step": 5751 + }, + { + "epoch": 79.8943231441048, + "grad_norm": 0.01626049168407917, + "learning_rate": 0.0006, + "loss": 4.207252502441406, + "step": 5752 + }, + { + "epoch": 79.90829694323143, + "grad_norm": 0.015415354631841183, + "learning_rate": 0.0006, + "loss": 4.151934623718262, + "step": 5753 + }, + { + "epoch": 79.92227074235808, + "grad_norm": 0.014158242382109165, + "learning_rate": 0.0006, + "loss": 4.3461995124816895, + "step": 5754 + }, + { + "epoch": 79.93624454148471, + "grad_norm": 0.01585628278553486, + "learning_rate": 0.0006, + "loss": 4.334761619567871, + "step": 5755 + }, + { + "epoch": 79.95021834061136, + "grad_norm": 0.013710759580135345, + "learning_rate": 0.0006, + "loss": 4.180401802062988, + "step": 5756 + }, + { + "epoch": 79.96419213973799, + "grad_norm": 0.014159204438328743, + "learning_rate": 0.0006, + "loss": 4.322331428527832, + "step": 5757 + }, + { + "epoch": 79.97816593886463, + "grad_norm": 0.015153195708990097, + "learning_rate": 0.0006, + "loss": 4.287478923797607, + "step": 5758 + }, + { + "epoch": 79.99213973799127, + "grad_norm": 0.014100227504968643, + "learning_rate": 0.0006, + "loss": 4.231807708740234, + "step": 5759 + }, + { + "epoch": 80.0, + "grad_norm": 0.015248682349920273, + "learning_rate": 0.0006, + "loss": 4.261106014251709, + "step": 5760 + }, + { + "epoch": 80.0, + "eval_loss": 4.610264301300049, + "eval_runtime": 56.514, + "eval_samples_per_second": 43.211, + "eval_steps_per_second": 1.362, + "step": 5760 + }, + { + "epoch": 80.01397379912663, + "grad_norm": 0.014730271883308887, + "learning_rate": 0.0006, + "loss": 4.108556747436523, + "step": 5761 + }, + { + "epoch": 80.02794759825328, + "grad_norm": 0.014575716108083725, + "learning_rate": 0.0006, + "loss": 4.169658660888672, + "step": 5762 + }, + { + "epoch": 80.04192139737991, + "grad_norm": 0.014706958085298538, + "learning_rate": 0.0006, + "loss": 4.252662658691406, + "step": 5763 + }, + { + "epoch": 80.05589519650655, + "grad_norm": 0.01561224739998579, + "learning_rate": 0.0006, + "loss": 4.193892955780029, + "step": 5764 + }, + { + "epoch": 80.06986899563319, + "grad_norm": 0.014702324755489826, + "learning_rate": 0.0006, + "loss": 4.1889238357543945, + "step": 5765 + }, + { + "epoch": 80.08384279475983, + "grad_norm": 0.015565741807222366, + "learning_rate": 0.0006, + "loss": 4.243549823760986, + "step": 5766 + }, + { + "epoch": 80.09781659388646, + "grad_norm": 0.01647181063890457, + "learning_rate": 0.0006, + "loss": 4.275446891784668, + "step": 5767 + }, + { + "epoch": 80.1117903930131, + "grad_norm": 0.018177716061472893, + "learning_rate": 0.0006, + "loss": 4.12040376663208, + "step": 5768 + }, + { + "epoch": 80.12576419213974, + "grad_norm": 0.01745191030204296, + "learning_rate": 0.0006, + "loss": 4.140371322631836, + "step": 5769 + }, + { + "epoch": 80.13973799126637, + "grad_norm": 0.01568993367254734, + "learning_rate": 0.0006, + "loss": 4.258280277252197, + "step": 5770 + }, + { + "epoch": 80.15371179039302, + "grad_norm": 0.01636597141623497, + "learning_rate": 0.0006, + "loss": 4.179052352905273, + "step": 5771 + }, + { + "epoch": 80.16768558951965, + "grad_norm": 0.016839100047945976, + "learning_rate": 0.0006, + "loss": 4.127636432647705, + "step": 5772 + }, + { + "epoch": 80.18165938864628, + "grad_norm": 0.015471016988158226, + "learning_rate": 0.0006, + "loss": 4.181325912475586, + "step": 5773 + }, + { + "epoch": 80.19563318777293, + "grad_norm": 0.015836089849472046, + "learning_rate": 0.0006, + "loss": 4.142168998718262, + "step": 5774 + }, + { + "epoch": 80.20960698689956, + "grad_norm": 0.014782954007387161, + "learning_rate": 0.0006, + "loss": 4.308799743652344, + "step": 5775 + }, + { + "epoch": 80.2235807860262, + "grad_norm": 0.014040893875062466, + "learning_rate": 0.0006, + "loss": 4.216936111450195, + "step": 5776 + }, + { + "epoch": 80.23755458515284, + "grad_norm": 0.01420280709862709, + "learning_rate": 0.0006, + "loss": 4.235230445861816, + "step": 5777 + }, + { + "epoch": 80.25152838427948, + "grad_norm": 0.013341007754206657, + "learning_rate": 0.0006, + "loss": 4.369109153747559, + "step": 5778 + }, + { + "epoch": 80.26550218340611, + "grad_norm": 0.014208327978849411, + "learning_rate": 0.0006, + "loss": 4.219112396240234, + "step": 5779 + }, + { + "epoch": 80.27947598253274, + "grad_norm": 0.013825052417814732, + "learning_rate": 0.0006, + "loss": 4.1983489990234375, + "step": 5780 + }, + { + "epoch": 80.29344978165939, + "grad_norm": 0.012916515581309795, + "learning_rate": 0.0006, + "loss": 4.204582691192627, + "step": 5781 + }, + { + "epoch": 80.30742358078602, + "grad_norm": 0.013071243651211262, + "learning_rate": 0.0006, + "loss": 4.17643928527832, + "step": 5782 + }, + { + "epoch": 80.32139737991267, + "grad_norm": 0.015507174655795097, + "learning_rate": 0.0006, + "loss": 4.172976016998291, + "step": 5783 + }, + { + "epoch": 80.3353711790393, + "grad_norm": 0.014967243187129498, + "learning_rate": 0.0006, + "loss": 4.141168594360352, + "step": 5784 + }, + { + "epoch": 80.34934497816595, + "grad_norm": 0.015629924833774567, + "learning_rate": 0.0006, + "loss": 4.253526210784912, + "step": 5785 + }, + { + "epoch": 80.36331877729258, + "grad_norm": 0.01607954502105713, + "learning_rate": 0.0006, + "loss": 4.249395370483398, + "step": 5786 + }, + { + "epoch": 80.37729257641921, + "grad_norm": 0.01500706560909748, + "learning_rate": 0.0006, + "loss": 4.279012680053711, + "step": 5787 + }, + { + "epoch": 80.39126637554585, + "grad_norm": 0.014343900606036186, + "learning_rate": 0.0006, + "loss": 4.268509864807129, + "step": 5788 + }, + { + "epoch": 80.40524017467249, + "grad_norm": 0.014682197012007236, + "learning_rate": 0.0006, + "loss": 4.206334114074707, + "step": 5789 + }, + { + "epoch": 80.41921397379913, + "grad_norm": 0.014989163726568222, + "learning_rate": 0.0006, + "loss": 4.143601417541504, + "step": 5790 + }, + { + "epoch": 80.43318777292576, + "grad_norm": 0.015001608058810234, + "learning_rate": 0.0006, + "loss": 4.164111137390137, + "step": 5791 + }, + { + "epoch": 80.44716157205241, + "grad_norm": 0.015222722664475441, + "learning_rate": 0.0006, + "loss": 4.2369914054870605, + "step": 5792 + }, + { + "epoch": 80.46113537117904, + "grad_norm": 0.014126183465123177, + "learning_rate": 0.0006, + "loss": 4.2209930419921875, + "step": 5793 + }, + { + "epoch": 80.47510917030567, + "grad_norm": 0.013065959326922894, + "learning_rate": 0.0006, + "loss": 4.231184959411621, + "step": 5794 + }, + { + "epoch": 80.48908296943232, + "grad_norm": 0.013833962380886078, + "learning_rate": 0.0006, + "loss": 4.226434707641602, + "step": 5795 + }, + { + "epoch": 80.50305676855895, + "grad_norm": 0.01387125626206398, + "learning_rate": 0.0006, + "loss": 4.241328716278076, + "step": 5796 + }, + { + "epoch": 80.5170305676856, + "grad_norm": 0.014145063236355782, + "learning_rate": 0.0006, + "loss": 4.2268805503845215, + "step": 5797 + }, + { + "epoch": 80.53100436681223, + "grad_norm": 0.012990240007638931, + "learning_rate": 0.0006, + "loss": 4.1718339920043945, + "step": 5798 + }, + { + "epoch": 80.54497816593886, + "grad_norm": 0.014185411855578423, + "learning_rate": 0.0006, + "loss": 4.232397556304932, + "step": 5799 + }, + { + "epoch": 80.5589519650655, + "grad_norm": 0.014260523952543736, + "learning_rate": 0.0006, + "loss": 4.282734394073486, + "step": 5800 + }, + { + "epoch": 80.57292576419214, + "grad_norm": 0.012542120181024075, + "learning_rate": 0.0006, + "loss": 4.134551525115967, + "step": 5801 + }, + { + "epoch": 80.58689956331878, + "grad_norm": 0.013078657910227776, + "learning_rate": 0.0006, + "loss": 4.229815483093262, + "step": 5802 + }, + { + "epoch": 80.60087336244541, + "grad_norm": 0.012821056880056858, + "learning_rate": 0.0006, + "loss": 4.208278179168701, + "step": 5803 + }, + { + "epoch": 80.61484716157206, + "grad_norm": 0.012964180670678616, + "learning_rate": 0.0006, + "loss": 4.188477039337158, + "step": 5804 + }, + { + "epoch": 80.62882096069869, + "grad_norm": 0.01631746254861355, + "learning_rate": 0.0006, + "loss": 4.3214497566223145, + "step": 5805 + }, + { + "epoch": 80.64279475982532, + "grad_norm": 0.01726640947163105, + "learning_rate": 0.0006, + "loss": 4.279359817504883, + "step": 5806 + }, + { + "epoch": 80.65676855895197, + "grad_norm": 0.018778258934617043, + "learning_rate": 0.0006, + "loss": 4.294047832489014, + "step": 5807 + }, + { + "epoch": 80.6707423580786, + "grad_norm": 0.01796831749379635, + "learning_rate": 0.0006, + "loss": 4.314162254333496, + "step": 5808 + }, + { + "epoch": 80.68471615720524, + "grad_norm": 0.016495538875460625, + "learning_rate": 0.0006, + "loss": 4.273871421813965, + "step": 5809 + }, + { + "epoch": 80.69868995633188, + "grad_norm": 0.013590306043624878, + "learning_rate": 0.0006, + "loss": 4.14961051940918, + "step": 5810 + }, + { + "epoch": 80.71266375545852, + "grad_norm": 0.014670426957309246, + "learning_rate": 0.0006, + "loss": 4.211853981018066, + "step": 5811 + }, + { + "epoch": 80.72663755458515, + "grad_norm": 0.01641341857612133, + "learning_rate": 0.0006, + "loss": 4.211948871612549, + "step": 5812 + }, + { + "epoch": 80.74061135371178, + "grad_norm": 0.016365470364689827, + "learning_rate": 0.0006, + "loss": 4.200774192810059, + "step": 5813 + }, + { + "epoch": 80.75458515283843, + "grad_norm": 0.01613970287144184, + "learning_rate": 0.0006, + "loss": 4.288322448730469, + "step": 5814 + }, + { + "epoch": 80.76855895196506, + "grad_norm": 0.016742777079343796, + "learning_rate": 0.0006, + "loss": 4.23801326751709, + "step": 5815 + }, + { + "epoch": 80.78253275109171, + "grad_norm": 0.015351069159805775, + "learning_rate": 0.0006, + "loss": 4.263832092285156, + "step": 5816 + }, + { + "epoch": 80.79650655021834, + "grad_norm": 0.013435336761176586, + "learning_rate": 0.0006, + "loss": 4.160857200622559, + "step": 5817 + }, + { + "epoch": 80.81048034934499, + "grad_norm": 0.01474886853247881, + "learning_rate": 0.0006, + "loss": 4.216025352478027, + "step": 5818 + }, + { + "epoch": 80.82445414847162, + "grad_norm": 0.01586935855448246, + "learning_rate": 0.0006, + "loss": 4.256402969360352, + "step": 5819 + }, + { + "epoch": 80.83842794759825, + "grad_norm": 0.014596572145819664, + "learning_rate": 0.0006, + "loss": 4.142584800720215, + "step": 5820 + }, + { + "epoch": 80.8524017467249, + "grad_norm": 0.01501593366265297, + "learning_rate": 0.0006, + "loss": 4.250462055206299, + "step": 5821 + }, + { + "epoch": 80.86637554585153, + "grad_norm": 0.014773511327803135, + "learning_rate": 0.0006, + "loss": 4.253305912017822, + "step": 5822 + }, + { + "epoch": 80.88034934497817, + "grad_norm": 0.013246326707303524, + "learning_rate": 0.0006, + "loss": 4.170881271362305, + "step": 5823 + }, + { + "epoch": 80.8943231441048, + "grad_norm": 0.013013756833970547, + "learning_rate": 0.0006, + "loss": 4.264721870422363, + "step": 5824 + }, + { + "epoch": 80.90829694323143, + "grad_norm": 0.013155903667211533, + "learning_rate": 0.0006, + "loss": 4.274667739868164, + "step": 5825 + }, + { + "epoch": 80.92227074235808, + "grad_norm": 0.01448952779173851, + "learning_rate": 0.0006, + "loss": 4.24777889251709, + "step": 5826 + }, + { + "epoch": 80.93624454148471, + "grad_norm": 0.01415231078863144, + "learning_rate": 0.0006, + "loss": 4.22014856338501, + "step": 5827 + }, + { + "epoch": 80.95021834061136, + "grad_norm": 0.014796273782849312, + "learning_rate": 0.0006, + "loss": 4.22036600112915, + "step": 5828 + }, + { + "epoch": 80.96419213973799, + "grad_norm": 0.013005326502025127, + "learning_rate": 0.0006, + "loss": 4.276123046875, + "step": 5829 + }, + { + "epoch": 80.97816593886463, + "grad_norm": 0.012851177714765072, + "learning_rate": 0.0006, + "loss": 4.248575687408447, + "step": 5830 + }, + { + "epoch": 80.99213973799127, + "grad_norm": 0.013740002177655697, + "learning_rate": 0.0006, + "loss": 4.051140308380127, + "step": 5831 + }, + { + "epoch": 81.0, + "grad_norm": 0.015296104364097118, + "learning_rate": 0.0006, + "loss": 4.251155853271484, + "step": 5832 + }, + { + "epoch": 81.0, + "eval_loss": 4.594127655029297, + "eval_runtime": 56.0959, + "eval_samples_per_second": 43.533, + "eval_steps_per_second": 1.373, + "step": 5832 + }, + { + "epoch": 81.01397379912663, + "grad_norm": 0.015804558992385864, + "learning_rate": 0.0006, + "loss": 4.152894020080566, + "step": 5833 + }, + { + "epoch": 81.02794759825328, + "grad_norm": 0.017000854015350342, + "learning_rate": 0.0006, + "loss": 4.266067981719971, + "step": 5834 + }, + { + "epoch": 81.04192139737991, + "grad_norm": 0.016351008787751198, + "learning_rate": 0.0006, + "loss": 4.192703723907471, + "step": 5835 + }, + { + "epoch": 81.05589519650655, + "grad_norm": 0.014703895896673203, + "learning_rate": 0.0006, + "loss": 4.23554801940918, + "step": 5836 + }, + { + "epoch": 81.06986899563319, + "grad_norm": 0.01619596965610981, + "learning_rate": 0.0006, + "loss": 4.233000755310059, + "step": 5837 + }, + { + "epoch": 81.08384279475983, + "grad_norm": 0.01815493032336235, + "learning_rate": 0.0006, + "loss": 4.200490951538086, + "step": 5838 + }, + { + "epoch": 81.09781659388646, + "grad_norm": 0.018305106088519096, + "learning_rate": 0.0006, + "loss": 4.154242515563965, + "step": 5839 + }, + { + "epoch": 81.1117903930131, + "grad_norm": 0.0200856514275074, + "learning_rate": 0.0006, + "loss": 4.25820779800415, + "step": 5840 + }, + { + "epoch": 81.12576419213974, + "grad_norm": 0.02101440541446209, + "learning_rate": 0.0006, + "loss": 4.168209075927734, + "step": 5841 + }, + { + "epoch": 81.13973799126637, + "grad_norm": 0.022829681634902954, + "learning_rate": 0.0006, + "loss": 4.25560998916626, + "step": 5842 + }, + { + "epoch": 81.15371179039302, + "grad_norm": 0.0195082426071167, + "learning_rate": 0.0006, + "loss": 4.2035627365112305, + "step": 5843 + }, + { + "epoch": 81.16768558951965, + "grad_norm": 0.01716248318552971, + "learning_rate": 0.0006, + "loss": 4.11322021484375, + "step": 5844 + }, + { + "epoch": 81.18165938864628, + "grad_norm": 0.02034664712846279, + "learning_rate": 0.0006, + "loss": 4.179389953613281, + "step": 5845 + }, + { + "epoch": 81.19563318777293, + "grad_norm": 0.023578710854053497, + "learning_rate": 0.0006, + "loss": 4.265695571899414, + "step": 5846 + }, + { + "epoch": 81.20960698689956, + "grad_norm": 0.019244182854890823, + "learning_rate": 0.0006, + "loss": 4.292043209075928, + "step": 5847 + }, + { + "epoch": 81.2235807860262, + "grad_norm": 0.01577291637659073, + "learning_rate": 0.0006, + "loss": 4.250776290893555, + "step": 5848 + }, + { + "epoch": 81.23755458515284, + "grad_norm": 0.01617090217769146, + "learning_rate": 0.0006, + "loss": 4.193098068237305, + "step": 5849 + }, + { + "epoch": 81.25152838427948, + "grad_norm": 0.016090184450149536, + "learning_rate": 0.0006, + "loss": 4.112357139587402, + "step": 5850 + }, + { + "epoch": 81.26550218340611, + "grad_norm": 0.01441365759819746, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 5851 + }, + { + "epoch": 81.27947598253274, + "grad_norm": 0.014122162945568562, + "learning_rate": 0.0006, + "loss": 4.276175022125244, + "step": 5852 + }, + { + "epoch": 81.29344978165939, + "grad_norm": 0.014859487302601337, + "learning_rate": 0.0006, + "loss": 4.094699859619141, + "step": 5853 + }, + { + "epoch": 81.30742358078602, + "grad_norm": 0.013624267652630806, + "learning_rate": 0.0006, + "loss": 4.2648773193359375, + "step": 5854 + }, + { + "epoch": 81.32139737991267, + "grad_norm": 0.01417913381010294, + "learning_rate": 0.0006, + "loss": 4.228567123413086, + "step": 5855 + }, + { + "epoch": 81.3353711790393, + "grad_norm": 0.014560805633664131, + "learning_rate": 0.0006, + "loss": 4.148930549621582, + "step": 5856 + }, + { + "epoch": 81.34934497816595, + "grad_norm": 0.013662154786288738, + "learning_rate": 0.0006, + "loss": 4.192107677459717, + "step": 5857 + }, + { + "epoch": 81.36331877729258, + "grad_norm": 0.013434980995953083, + "learning_rate": 0.0006, + "loss": 4.162052154541016, + "step": 5858 + }, + { + "epoch": 81.37729257641921, + "grad_norm": 0.01300742756575346, + "learning_rate": 0.0006, + "loss": 4.262385368347168, + "step": 5859 + }, + { + "epoch": 81.39126637554585, + "grad_norm": 0.012544351629912853, + "learning_rate": 0.0006, + "loss": 4.215371131896973, + "step": 5860 + }, + { + "epoch": 81.40524017467249, + "grad_norm": 0.013663645833730698, + "learning_rate": 0.0006, + "loss": 4.247312545776367, + "step": 5861 + }, + { + "epoch": 81.41921397379913, + "grad_norm": 0.013127296231687069, + "learning_rate": 0.0006, + "loss": 4.087685585021973, + "step": 5862 + }, + { + "epoch": 81.43318777292576, + "grad_norm": 0.014081938192248344, + "learning_rate": 0.0006, + "loss": 4.272648811340332, + "step": 5863 + }, + { + "epoch": 81.44716157205241, + "grad_norm": 0.015470641665160656, + "learning_rate": 0.0006, + "loss": 4.278578758239746, + "step": 5864 + }, + { + "epoch": 81.46113537117904, + "grad_norm": 0.016524532809853554, + "learning_rate": 0.0006, + "loss": 4.118268013000488, + "step": 5865 + }, + { + "epoch": 81.47510917030567, + "grad_norm": 0.016242722049355507, + "learning_rate": 0.0006, + "loss": 4.182405471801758, + "step": 5866 + }, + { + "epoch": 81.48908296943232, + "grad_norm": 0.01557017583400011, + "learning_rate": 0.0006, + "loss": 4.229862213134766, + "step": 5867 + }, + { + "epoch": 81.50305676855895, + "grad_norm": 0.01562870666384697, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 5868 + }, + { + "epoch": 81.5170305676856, + "grad_norm": 0.018606074154376984, + "learning_rate": 0.0006, + "loss": 4.148681640625, + "step": 5869 + }, + { + "epoch": 81.53100436681223, + "grad_norm": 0.017740730196237564, + "learning_rate": 0.0006, + "loss": 4.321932792663574, + "step": 5870 + }, + { + "epoch": 81.54497816593886, + "grad_norm": 0.014517677947878838, + "learning_rate": 0.0006, + "loss": 4.22862434387207, + "step": 5871 + }, + { + "epoch": 81.5589519650655, + "grad_norm": 0.01389714702963829, + "learning_rate": 0.0006, + "loss": 4.198000907897949, + "step": 5872 + }, + { + "epoch": 81.57292576419214, + "grad_norm": 0.015348682180047035, + "learning_rate": 0.0006, + "loss": 4.293485164642334, + "step": 5873 + }, + { + "epoch": 81.58689956331878, + "grad_norm": 0.015965791419148445, + "learning_rate": 0.0006, + "loss": 4.23481559753418, + "step": 5874 + }, + { + "epoch": 81.60087336244541, + "grad_norm": 0.015662726014852524, + "learning_rate": 0.0006, + "loss": 4.158148765563965, + "step": 5875 + }, + { + "epoch": 81.61484716157206, + "grad_norm": 0.014402080327272415, + "learning_rate": 0.0006, + "loss": 4.132723808288574, + "step": 5876 + }, + { + "epoch": 81.62882096069869, + "grad_norm": 0.016846125945448875, + "learning_rate": 0.0006, + "loss": 4.1654462814331055, + "step": 5877 + }, + { + "epoch": 81.64279475982532, + "grad_norm": 0.0196097232401371, + "learning_rate": 0.0006, + "loss": 4.228450298309326, + "step": 5878 + }, + { + "epoch": 81.65676855895197, + "grad_norm": 0.016005294397473335, + "learning_rate": 0.0006, + "loss": 4.230222225189209, + "step": 5879 + }, + { + "epoch": 81.6707423580786, + "grad_norm": 0.014685059897601604, + "learning_rate": 0.0006, + "loss": 4.1367387771606445, + "step": 5880 + }, + { + "epoch": 81.68471615720524, + "grad_norm": 0.01647144928574562, + "learning_rate": 0.0006, + "loss": 4.096675395965576, + "step": 5881 + }, + { + "epoch": 81.69868995633188, + "grad_norm": 0.015978610143065453, + "learning_rate": 0.0006, + "loss": 4.232292175292969, + "step": 5882 + }, + { + "epoch": 81.71266375545852, + "grad_norm": 0.01510288193821907, + "learning_rate": 0.0006, + "loss": 4.274591445922852, + "step": 5883 + }, + { + "epoch": 81.72663755458515, + "grad_norm": 0.014942476525902748, + "learning_rate": 0.0006, + "loss": 4.161482810974121, + "step": 5884 + }, + { + "epoch": 81.74061135371178, + "grad_norm": 0.014595299027860165, + "learning_rate": 0.0006, + "loss": 4.204291343688965, + "step": 5885 + }, + { + "epoch": 81.75458515283843, + "grad_norm": 0.014952429570257664, + "learning_rate": 0.0006, + "loss": 4.242499351501465, + "step": 5886 + }, + { + "epoch": 81.76855895196506, + "grad_norm": 0.014761554077267647, + "learning_rate": 0.0006, + "loss": 4.20393180847168, + "step": 5887 + }, + { + "epoch": 81.78253275109171, + "grad_norm": 0.014749647118151188, + "learning_rate": 0.0006, + "loss": 4.227565765380859, + "step": 5888 + }, + { + "epoch": 81.79650655021834, + "grad_norm": 0.014384493231773376, + "learning_rate": 0.0006, + "loss": 4.282228946685791, + "step": 5889 + }, + { + "epoch": 81.81048034934499, + "grad_norm": 0.01260330155491829, + "learning_rate": 0.0006, + "loss": 4.2882866859436035, + "step": 5890 + }, + { + "epoch": 81.82445414847162, + "grad_norm": 0.014055311679840088, + "learning_rate": 0.0006, + "loss": 4.22351598739624, + "step": 5891 + }, + { + "epoch": 81.83842794759825, + "grad_norm": 0.013674065470695496, + "learning_rate": 0.0006, + "loss": 4.240716934204102, + "step": 5892 + }, + { + "epoch": 81.8524017467249, + "grad_norm": 0.01225196197628975, + "learning_rate": 0.0006, + "loss": 4.2958598136901855, + "step": 5893 + }, + { + "epoch": 81.86637554585153, + "grad_norm": 0.013538680970668793, + "learning_rate": 0.0006, + "loss": 4.24342679977417, + "step": 5894 + }, + { + "epoch": 81.88034934497817, + "grad_norm": 0.013151394203305244, + "learning_rate": 0.0006, + "loss": 4.081999778747559, + "step": 5895 + }, + { + "epoch": 81.8943231441048, + "grad_norm": 0.014260678552091122, + "learning_rate": 0.0006, + "loss": 4.156423568725586, + "step": 5896 + }, + { + "epoch": 81.90829694323143, + "grad_norm": 0.01329776830971241, + "learning_rate": 0.0006, + "loss": 4.1713995933532715, + "step": 5897 + }, + { + "epoch": 81.92227074235808, + "grad_norm": 0.014623276889324188, + "learning_rate": 0.0006, + "loss": 4.244891166687012, + "step": 5898 + }, + { + "epoch": 81.93624454148471, + "grad_norm": 0.015760324895381927, + "learning_rate": 0.0006, + "loss": 4.076528549194336, + "step": 5899 + }, + { + "epoch": 81.95021834061136, + "grad_norm": 0.014331766404211521, + "learning_rate": 0.0006, + "loss": 4.283037185668945, + "step": 5900 + }, + { + "epoch": 81.96419213973799, + "grad_norm": 0.012383842840790749, + "learning_rate": 0.0006, + "loss": 4.117308616638184, + "step": 5901 + }, + { + "epoch": 81.97816593886463, + "grad_norm": 0.01250660140067339, + "learning_rate": 0.0006, + "loss": 4.241250991821289, + "step": 5902 + }, + { + "epoch": 81.99213973799127, + "grad_norm": 0.012276900932192802, + "learning_rate": 0.0006, + "loss": 4.141091823577881, + "step": 5903 + }, + { + "epoch": 82.0, + "grad_norm": 0.01588737778365612, + "learning_rate": 0.0006, + "loss": 4.065532684326172, + "step": 5904 + }, + { + "epoch": 82.0, + "eval_loss": 4.612304210662842, + "eval_runtime": 57.5411, + "eval_samples_per_second": 42.439, + "eval_steps_per_second": 1.338, + "step": 5904 + }, + { + "epoch": 82.01397379912663, + "grad_norm": 0.017433544620871544, + "learning_rate": 0.0006, + "loss": 4.202497482299805, + "step": 5905 + }, + { + "epoch": 82.02794759825328, + "grad_norm": 0.017260415479540825, + "learning_rate": 0.0006, + "loss": 4.099949359893799, + "step": 5906 + }, + { + "epoch": 82.04192139737991, + "grad_norm": 0.01577908545732498, + "learning_rate": 0.0006, + "loss": 4.17806339263916, + "step": 5907 + }, + { + "epoch": 82.05589519650655, + "grad_norm": 0.016093818470835686, + "learning_rate": 0.0006, + "loss": 4.219491004943848, + "step": 5908 + }, + { + "epoch": 82.06986899563319, + "grad_norm": 0.015426302328705788, + "learning_rate": 0.0006, + "loss": 4.207982063293457, + "step": 5909 + }, + { + "epoch": 82.08384279475983, + "grad_norm": 0.01605442352592945, + "learning_rate": 0.0006, + "loss": 4.265942573547363, + "step": 5910 + }, + { + "epoch": 82.09781659388646, + "grad_norm": 0.01787225715816021, + "learning_rate": 0.0006, + "loss": 4.151535511016846, + "step": 5911 + }, + { + "epoch": 82.1117903930131, + "grad_norm": 0.01750164106488228, + "learning_rate": 0.0006, + "loss": 4.133655071258545, + "step": 5912 + }, + { + "epoch": 82.12576419213974, + "grad_norm": 0.01602957211434841, + "learning_rate": 0.0006, + "loss": 4.176972389221191, + "step": 5913 + }, + { + "epoch": 82.13973799126637, + "grad_norm": 0.01553812064230442, + "learning_rate": 0.0006, + "loss": 4.125692367553711, + "step": 5914 + }, + { + "epoch": 82.15371179039302, + "grad_norm": 0.013978508301079273, + "learning_rate": 0.0006, + "loss": 4.207271575927734, + "step": 5915 + }, + { + "epoch": 82.16768558951965, + "grad_norm": 0.013561118394136429, + "learning_rate": 0.0006, + "loss": 4.238930702209473, + "step": 5916 + }, + { + "epoch": 82.18165938864628, + "grad_norm": 0.013285507448017597, + "learning_rate": 0.0006, + "loss": 4.129464626312256, + "step": 5917 + }, + { + "epoch": 82.19563318777293, + "grad_norm": 0.013903839513659477, + "learning_rate": 0.0006, + "loss": 4.214109420776367, + "step": 5918 + }, + { + "epoch": 82.20960698689956, + "grad_norm": 0.014426837675273418, + "learning_rate": 0.0006, + "loss": 4.143196105957031, + "step": 5919 + }, + { + "epoch": 82.2235807860262, + "grad_norm": 0.015347165986895561, + "learning_rate": 0.0006, + "loss": 4.235439300537109, + "step": 5920 + }, + { + "epoch": 82.23755458515284, + "grad_norm": 0.014266039244830608, + "learning_rate": 0.0006, + "loss": 3.991617202758789, + "step": 5921 + }, + { + "epoch": 82.25152838427948, + "grad_norm": 0.013101593591272831, + "learning_rate": 0.0006, + "loss": 4.266017913818359, + "step": 5922 + }, + { + "epoch": 82.26550218340611, + "grad_norm": 0.013741881586611271, + "learning_rate": 0.0006, + "loss": 4.224879741668701, + "step": 5923 + }, + { + "epoch": 82.27947598253274, + "grad_norm": 0.014206945896148682, + "learning_rate": 0.0006, + "loss": 4.102272987365723, + "step": 5924 + }, + { + "epoch": 82.29344978165939, + "grad_norm": 0.013798723928630352, + "learning_rate": 0.0006, + "loss": 4.30418586730957, + "step": 5925 + }, + { + "epoch": 82.30742358078602, + "grad_norm": 0.015470122918486595, + "learning_rate": 0.0006, + "loss": 4.2513322830200195, + "step": 5926 + }, + { + "epoch": 82.32139737991267, + "grad_norm": 0.01655442826449871, + "learning_rate": 0.0006, + "loss": 4.223002910614014, + "step": 5927 + }, + { + "epoch": 82.3353711790393, + "grad_norm": 0.017170004546642303, + "learning_rate": 0.0006, + "loss": 4.253904819488525, + "step": 5928 + }, + { + "epoch": 82.34934497816595, + "grad_norm": 0.017595812678337097, + "learning_rate": 0.0006, + "loss": 4.184927940368652, + "step": 5929 + }, + { + "epoch": 82.36331877729258, + "grad_norm": 0.015073256567120552, + "learning_rate": 0.0006, + "loss": 4.202239036560059, + "step": 5930 + }, + { + "epoch": 82.37729257641921, + "grad_norm": 0.014310941100120544, + "learning_rate": 0.0006, + "loss": 4.124934673309326, + "step": 5931 + }, + { + "epoch": 82.39126637554585, + "grad_norm": 0.013671821914613247, + "learning_rate": 0.0006, + "loss": 4.138919353485107, + "step": 5932 + }, + { + "epoch": 82.40524017467249, + "grad_norm": 0.014103314839303493, + "learning_rate": 0.0006, + "loss": 4.211289405822754, + "step": 5933 + }, + { + "epoch": 82.41921397379913, + "grad_norm": 0.014272668398916721, + "learning_rate": 0.0006, + "loss": 4.06290340423584, + "step": 5934 + }, + { + "epoch": 82.43318777292576, + "grad_norm": 0.01387016475200653, + "learning_rate": 0.0006, + "loss": 4.101334571838379, + "step": 5935 + }, + { + "epoch": 82.44716157205241, + "grad_norm": 0.013318683952093124, + "learning_rate": 0.0006, + "loss": 4.206409931182861, + "step": 5936 + }, + { + "epoch": 82.46113537117904, + "grad_norm": 0.013781268149614334, + "learning_rate": 0.0006, + "loss": 4.189474105834961, + "step": 5937 + }, + { + "epoch": 82.47510917030567, + "grad_norm": 0.013195851817727089, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 5938 + }, + { + "epoch": 82.48908296943232, + "grad_norm": 0.014358163811266422, + "learning_rate": 0.0006, + "loss": 4.203262805938721, + "step": 5939 + }, + { + "epoch": 82.50305676855895, + "grad_norm": 0.01403636671602726, + "learning_rate": 0.0006, + "loss": 4.079258441925049, + "step": 5940 + }, + { + "epoch": 82.5170305676856, + "grad_norm": 0.013272535055875778, + "learning_rate": 0.0006, + "loss": 4.19197940826416, + "step": 5941 + }, + { + "epoch": 82.53100436681223, + "grad_norm": 0.013787264935672283, + "learning_rate": 0.0006, + "loss": 4.12995719909668, + "step": 5942 + }, + { + "epoch": 82.54497816593886, + "grad_norm": 0.015020057559013367, + "learning_rate": 0.0006, + "loss": 4.148890495300293, + "step": 5943 + }, + { + "epoch": 82.5589519650655, + "grad_norm": 0.014199157245457172, + "learning_rate": 0.0006, + "loss": 4.201516151428223, + "step": 5944 + }, + { + "epoch": 82.57292576419214, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.240894317626953, + "step": 5945 + }, + { + "epoch": 82.58689956331878, + "grad_norm": 0.014767560176551342, + "learning_rate": 0.0006, + "loss": 4.173320770263672, + "step": 5946 + }, + { + "epoch": 82.60087336244541, + "grad_norm": 0.01612078957259655, + "learning_rate": 0.0006, + "loss": 4.266448020935059, + "step": 5947 + }, + { + "epoch": 82.61484716157206, + "grad_norm": 0.016706543043255806, + "learning_rate": 0.0006, + "loss": 4.229398250579834, + "step": 5948 + }, + { + "epoch": 82.62882096069869, + "grad_norm": 0.015630796551704407, + "learning_rate": 0.0006, + "loss": 4.181282997131348, + "step": 5949 + }, + { + "epoch": 82.64279475982532, + "grad_norm": 0.013103844597935677, + "learning_rate": 0.0006, + "loss": 4.309617042541504, + "step": 5950 + }, + { + "epoch": 82.65676855895197, + "grad_norm": 0.016136541962623596, + "learning_rate": 0.0006, + "loss": 4.224159240722656, + "step": 5951 + }, + { + "epoch": 82.6707423580786, + "grad_norm": 0.016160814091563225, + "learning_rate": 0.0006, + "loss": 4.267970085144043, + "step": 5952 + }, + { + "epoch": 82.68471615720524, + "grad_norm": 0.01672680303454399, + "learning_rate": 0.0006, + "loss": 4.230506896972656, + "step": 5953 + }, + { + "epoch": 82.69868995633188, + "grad_norm": 0.0159586351364851, + "learning_rate": 0.0006, + "loss": 4.159542083740234, + "step": 5954 + }, + { + "epoch": 82.71266375545852, + "grad_norm": 0.01461219135671854, + "learning_rate": 0.0006, + "loss": 4.27863073348999, + "step": 5955 + }, + { + "epoch": 82.72663755458515, + "grad_norm": 0.015088613145053387, + "learning_rate": 0.0006, + "loss": 4.272418975830078, + "step": 5956 + }, + { + "epoch": 82.74061135371178, + "grad_norm": 0.01594499684870243, + "learning_rate": 0.0006, + "loss": 4.271326065063477, + "step": 5957 + }, + { + "epoch": 82.75458515283843, + "grad_norm": 0.016648447141051292, + "learning_rate": 0.0006, + "loss": 4.141478538513184, + "step": 5958 + }, + { + "epoch": 82.76855895196506, + "grad_norm": 0.01540316641330719, + "learning_rate": 0.0006, + "loss": 4.111814498901367, + "step": 5959 + }, + { + "epoch": 82.78253275109171, + "grad_norm": 0.01634364202618599, + "learning_rate": 0.0006, + "loss": 4.3113112449646, + "step": 5960 + }, + { + "epoch": 82.79650655021834, + "grad_norm": 0.014768977649509907, + "learning_rate": 0.0006, + "loss": 4.184217929840088, + "step": 5961 + }, + { + "epoch": 82.81048034934499, + "grad_norm": 0.013453755527734756, + "learning_rate": 0.0006, + "loss": 4.247729301452637, + "step": 5962 + }, + { + "epoch": 82.82445414847162, + "grad_norm": 0.014731907285749912, + "learning_rate": 0.0006, + "loss": 4.352839469909668, + "step": 5963 + }, + { + "epoch": 82.83842794759825, + "grad_norm": 0.016047345474362373, + "learning_rate": 0.0006, + "loss": 4.257283687591553, + "step": 5964 + }, + { + "epoch": 82.8524017467249, + "grad_norm": 0.017141170799732208, + "learning_rate": 0.0006, + "loss": 4.22850227355957, + "step": 5965 + }, + { + "epoch": 82.86637554585153, + "grad_norm": 0.01370396837592125, + "learning_rate": 0.0006, + "loss": 4.1169257164001465, + "step": 5966 + }, + { + "epoch": 82.88034934497817, + "grad_norm": 0.012060081586241722, + "learning_rate": 0.0006, + "loss": 4.195242404937744, + "step": 5967 + }, + { + "epoch": 82.8943231441048, + "grad_norm": 0.015528378076851368, + "learning_rate": 0.0006, + "loss": 4.294939994812012, + "step": 5968 + }, + { + "epoch": 82.90829694323143, + "grad_norm": 0.018151041120290756, + "learning_rate": 0.0006, + "loss": 4.137192726135254, + "step": 5969 + }, + { + "epoch": 82.92227074235808, + "grad_norm": 0.017895735800266266, + "learning_rate": 0.0006, + "loss": 4.174036026000977, + "step": 5970 + }, + { + "epoch": 82.93624454148471, + "grad_norm": 0.015610386617481709, + "learning_rate": 0.0006, + "loss": 4.230136871337891, + "step": 5971 + }, + { + "epoch": 82.95021834061136, + "grad_norm": 0.01339928898960352, + "learning_rate": 0.0006, + "loss": 4.199074745178223, + "step": 5972 + }, + { + "epoch": 82.96419213973799, + "grad_norm": 0.015454214997589588, + "learning_rate": 0.0006, + "loss": 4.216109752655029, + "step": 5973 + }, + { + "epoch": 82.97816593886463, + "grad_norm": 0.014978132210671902, + "learning_rate": 0.0006, + "loss": 4.131511688232422, + "step": 5974 + }, + { + "epoch": 82.99213973799127, + "grad_norm": 0.015611122362315655, + "learning_rate": 0.0006, + "loss": 4.246518135070801, + "step": 5975 + }, + { + "epoch": 83.0, + "grad_norm": 0.016682764515280724, + "learning_rate": 0.0006, + "loss": 4.110680103302002, + "step": 5976 + }, + { + "epoch": 83.0, + "eval_loss": 4.641707897186279, + "eval_runtime": 57.2799, + "eval_samples_per_second": 42.633, + "eval_steps_per_second": 1.344, + "step": 5976 + }, + { + "epoch": 83.01397379912663, + "grad_norm": 0.015007982030510902, + "learning_rate": 0.0006, + "loss": 4.100912094116211, + "step": 5977 + }, + { + "epoch": 83.02794759825328, + "grad_norm": 0.017609436064958572, + "learning_rate": 0.0006, + "loss": 4.233399868011475, + "step": 5978 + }, + { + "epoch": 83.04192139737991, + "grad_norm": 0.019817272201180458, + "learning_rate": 0.0006, + "loss": 4.191173076629639, + "step": 5979 + }, + { + "epoch": 83.05589519650655, + "grad_norm": 0.017752140760421753, + "learning_rate": 0.0006, + "loss": 4.044437885284424, + "step": 5980 + }, + { + "epoch": 83.06986899563319, + "grad_norm": 0.015482760034501553, + "learning_rate": 0.0006, + "loss": 4.319931983947754, + "step": 5981 + }, + { + "epoch": 83.08384279475983, + "grad_norm": 0.015153988264501095, + "learning_rate": 0.0006, + "loss": 4.095277309417725, + "step": 5982 + }, + { + "epoch": 83.09781659388646, + "grad_norm": 0.017025282606482506, + "learning_rate": 0.0006, + "loss": 4.235110759735107, + "step": 5983 + }, + { + "epoch": 83.1117903930131, + "grad_norm": 0.01566586270928383, + "learning_rate": 0.0006, + "loss": 4.1934685707092285, + "step": 5984 + }, + { + "epoch": 83.12576419213974, + "grad_norm": 0.016419263556599617, + "learning_rate": 0.0006, + "loss": 4.234419822692871, + "step": 5985 + }, + { + "epoch": 83.13973799126637, + "grad_norm": 0.017132675275206566, + "learning_rate": 0.0006, + "loss": 4.221447944641113, + "step": 5986 + }, + { + "epoch": 83.15371179039302, + "grad_norm": 0.016138629987835884, + "learning_rate": 0.0006, + "loss": 4.199830055236816, + "step": 5987 + }, + { + "epoch": 83.16768558951965, + "grad_norm": 0.016210174188017845, + "learning_rate": 0.0006, + "loss": 4.081717491149902, + "step": 5988 + }, + { + "epoch": 83.18165938864628, + "grad_norm": 0.016055308282375336, + "learning_rate": 0.0006, + "loss": 4.247655391693115, + "step": 5989 + }, + { + "epoch": 83.19563318777293, + "grad_norm": 0.016132952645421028, + "learning_rate": 0.0006, + "loss": 4.262535572052002, + "step": 5990 + }, + { + "epoch": 83.20960698689956, + "grad_norm": 0.01779935695230961, + "learning_rate": 0.0006, + "loss": 4.153120040893555, + "step": 5991 + }, + { + "epoch": 83.2235807860262, + "grad_norm": 0.018360646441578865, + "learning_rate": 0.0006, + "loss": 4.232644081115723, + "step": 5992 + }, + { + "epoch": 83.23755458515284, + "grad_norm": 0.019629675894975662, + "learning_rate": 0.0006, + "loss": 3.9758734703063965, + "step": 5993 + }, + { + "epoch": 83.25152838427948, + "grad_norm": 0.017819996923208237, + "learning_rate": 0.0006, + "loss": 4.196932315826416, + "step": 5994 + }, + { + "epoch": 83.26550218340611, + "grad_norm": 0.01557681243866682, + "learning_rate": 0.0006, + "loss": 4.122654914855957, + "step": 5995 + }, + { + "epoch": 83.27947598253274, + "grad_norm": 0.017858853563666344, + "learning_rate": 0.0006, + "loss": 4.185330867767334, + "step": 5996 + }, + { + "epoch": 83.29344978165939, + "grad_norm": 0.018065953627228737, + "learning_rate": 0.0006, + "loss": 4.218471050262451, + "step": 5997 + }, + { + "epoch": 83.30742358078602, + "grad_norm": 0.017247939482331276, + "learning_rate": 0.0006, + "loss": 4.127076148986816, + "step": 5998 + }, + { + "epoch": 83.32139737991267, + "grad_norm": 0.017047952860593796, + "learning_rate": 0.0006, + "loss": 4.131356716156006, + "step": 5999 + }, + { + "epoch": 83.3353711790393, + "grad_norm": 0.015905629843473434, + "learning_rate": 0.0006, + "loss": 4.2239789962768555, + "step": 6000 + }, + { + "epoch": 83.34934497816595, + "grad_norm": 0.016522347927093506, + "learning_rate": 0.0006, + "loss": 4.054897785186768, + "step": 6001 + }, + { + "epoch": 83.36331877729258, + "grad_norm": 0.01622571051120758, + "learning_rate": 0.0006, + "loss": 4.284210681915283, + "step": 6002 + }, + { + "epoch": 83.37729257641921, + "grad_norm": 0.01640467159450054, + "learning_rate": 0.0006, + "loss": 4.268840789794922, + "step": 6003 + }, + { + "epoch": 83.39126637554585, + "grad_norm": 0.015698859468102455, + "learning_rate": 0.0006, + "loss": 4.179793357849121, + "step": 6004 + }, + { + "epoch": 83.40524017467249, + "grad_norm": 0.016694528982043266, + "learning_rate": 0.0006, + "loss": 4.132297515869141, + "step": 6005 + }, + { + "epoch": 83.41921397379913, + "grad_norm": 0.015345900319516659, + "learning_rate": 0.0006, + "loss": 4.090163230895996, + "step": 6006 + }, + { + "epoch": 83.43318777292576, + "grad_norm": 0.014722758904099464, + "learning_rate": 0.0006, + "loss": 4.179166793823242, + "step": 6007 + }, + { + "epoch": 83.44716157205241, + "grad_norm": 0.014884887263178825, + "learning_rate": 0.0006, + "loss": 4.137659072875977, + "step": 6008 + }, + { + "epoch": 83.46113537117904, + "grad_norm": 0.014041902497410774, + "learning_rate": 0.0006, + "loss": 4.1202073097229, + "step": 6009 + }, + { + "epoch": 83.47510917030567, + "grad_norm": 0.012799333781003952, + "learning_rate": 0.0006, + "loss": 4.092607498168945, + "step": 6010 + }, + { + "epoch": 83.48908296943232, + "grad_norm": 0.013272707350552082, + "learning_rate": 0.0006, + "loss": 4.132715225219727, + "step": 6011 + }, + { + "epoch": 83.50305676855895, + "grad_norm": 0.012934371829032898, + "learning_rate": 0.0006, + "loss": 4.230278968811035, + "step": 6012 + }, + { + "epoch": 83.5170305676856, + "grad_norm": 0.0120006762444973, + "learning_rate": 0.0006, + "loss": 4.140068531036377, + "step": 6013 + }, + { + "epoch": 83.53100436681223, + "grad_norm": 0.013616700656712055, + "learning_rate": 0.0006, + "loss": 4.20107364654541, + "step": 6014 + }, + { + "epoch": 83.54497816593886, + "grad_norm": 0.013957851566374302, + "learning_rate": 0.0006, + "loss": 4.124945163726807, + "step": 6015 + }, + { + "epoch": 83.5589519650655, + "grad_norm": 0.012762055732309818, + "learning_rate": 0.0006, + "loss": 4.022052764892578, + "step": 6016 + }, + { + "epoch": 83.57292576419214, + "grad_norm": 0.013080576434731483, + "learning_rate": 0.0006, + "loss": 4.103252410888672, + "step": 6017 + }, + { + "epoch": 83.58689956331878, + "grad_norm": 0.013957500457763672, + "learning_rate": 0.0006, + "loss": 4.235089302062988, + "step": 6018 + }, + { + "epoch": 83.60087336244541, + "grad_norm": 0.014707164838910103, + "learning_rate": 0.0006, + "loss": 4.191570281982422, + "step": 6019 + }, + { + "epoch": 83.61484716157206, + "grad_norm": 0.012807869352400303, + "learning_rate": 0.0006, + "loss": 4.294543266296387, + "step": 6020 + }, + { + "epoch": 83.62882096069869, + "grad_norm": 0.012027994729578495, + "learning_rate": 0.0006, + "loss": 4.13849401473999, + "step": 6021 + }, + { + "epoch": 83.64279475982532, + "grad_norm": 0.012147724628448486, + "learning_rate": 0.0006, + "loss": 4.127481937408447, + "step": 6022 + }, + { + "epoch": 83.65676855895197, + "grad_norm": 0.013241392560303211, + "learning_rate": 0.0006, + "loss": 4.218972682952881, + "step": 6023 + }, + { + "epoch": 83.6707423580786, + "grad_norm": 0.014071648940443993, + "learning_rate": 0.0006, + "loss": 4.096724033355713, + "step": 6024 + }, + { + "epoch": 83.68471615720524, + "grad_norm": 0.015212797559797764, + "learning_rate": 0.0006, + "loss": 4.27097225189209, + "step": 6025 + }, + { + "epoch": 83.69868995633188, + "grad_norm": 0.017647385597229004, + "learning_rate": 0.0006, + "loss": 4.209835529327393, + "step": 6026 + }, + { + "epoch": 83.71266375545852, + "grad_norm": 0.01709647662937641, + "learning_rate": 0.0006, + "loss": 4.042106628417969, + "step": 6027 + }, + { + "epoch": 83.72663755458515, + "grad_norm": 0.014530927874147892, + "learning_rate": 0.0006, + "loss": 4.181731700897217, + "step": 6028 + }, + { + "epoch": 83.74061135371178, + "grad_norm": 0.015356711111962795, + "learning_rate": 0.0006, + "loss": 4.314183235168457, + "step": 6029 + }, + { + "epoch": 83.75458515283843, + "grad_norm": 0.015500029549002647, + "learning_rate": 0.0006, + "loss": 4.06776762008667, + "step": 6030 + }, + { + "epoch": 83.76855895196506, + "grad_norm": 0.015351332724094391, + "learning_rate": 0.0006, + "loss": 4.185141563415527, + "step": 6031 + }, + { + "epoch": 83.78253275109171, + "grad_norm": 0.016678154468536377, + "learning_rate": 0.0006, + "loss": 4.179529666900635, + "step": 6032 + }, + { + "epoch": 83.79650655021834, + "grad_norm": 0.0156265702098608, + "learning_rate": 0.0006, + "loss": 4.238584041595459, + "step": 6033 + }, + { + "epoch": 83.81048034934499, + "grad_norm": 0.016418596729636192, + "learning_rate": 0.0006, + "loss": 4.1593523025512695, + "step": 6034 + }, + { + "epoch": 83.82445414847162, + "grad_norm": 0.016113586723804474, + "learning_rate": 0.0006, + "loss": 4.157470226287842, + "step": 6035 + }, + { + "epoch": 83.83842794759825, + "grad_norm": 0.014434042386710644, + "learning_rate": 0.0006, + "loss": 4.196226119995117, + "step": 6036 + }, + { + "epoch": 83.8524017467249, + "grad_norm": 0.012647042982280254, + "learning_rate": 0.0006, + "loss": 4.130589485168457, + "step": 6037 + }, + { + "epoch": 83.86637554585153, + "grad_norm": 0.014412877149879932, + "learning_rate": 0.0006, + "loss": 4.154845237731934, + "step": 6038 + }, + { + "epoch": 83.88034934497817, + "grad_norm": 0.014222027733922005, + "learning_rate": 0.0006, + "loss": 4.189724922180176, + "step": 6039 + }, + { + "epoch": 83.8943231441048, + "grad_norm": 0.013084255158901215, + "learning_rate": 0.0006, + "loss": 4.303338527679443, + "step": 6040 + }, + { + "epoch": 83.90829694323143, + "grad_norm": 0.012659232132136822, + "learning_rate": 0.0006, + "loss": 4.08192253112793, + "step": 6041 + }, + { + "epoch": 83.92227074235808, + "grad_norm": 0.01246276218444109, + "learning_rate": 0.0006, + "loss": 4.283324718475342, + "step": 6042 + }, + { + "epoch": 83.93624454148471, + "grad_norm": 0.01340118795633316, + "learning_rate": 0.0006, + "loss": 4.117908477783203, + "step": 6043 + }, + { + "epoch": 83.95021834061136, + "grad_norm": 0.014496722258627415, + "learning_rate": 0.0006, + "loss": 4.110318660736084, + "step": 6044 + }, + { + "epoch": 83.96419213973799, + "grad_norm": 0.013348298147320747, + "learning_rate": 0.0006, + "loss": 4.185139179229736, + "step": 6045 + }, + { + "epoch": 83.97816593886463, + "grad_norm": 0.013294277712702751, + "learning_rate": 0.0006, + "loss": 4.271645545959473, + "step": 6046 + }, + { + "epoch": 83.99213973799127, + "grad_norm": 0.014431594870984554, + "learning_rate": 0.0006, + "loss": 4.161383152008057, + "step": 6047 + }, + { + "epoch": 84.0, + "grad_norm": 0.016268927603960037, + "learning_rate": 0.0006, + "loss": 4.178324222564697, + "step": 6048 + }, + { + "epoch": 84.0, + "eval_loss": 4.6307549476623535, + "eval_runtime": 56.8521, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 6048 + }, + { + "epoch": 84.01397379912663, + "grad_norm": 0.014864908531308174, + "learning_rate": 0.0006, + "loss": 4.121300220489502, + "step": 6049 + }, + { + "epoch": 84.02794759825328, + "grad_norm": 0.014863367192447186, + "learning_rate": 0.0006, + "loss": 4.1650543212890625, + "step": 6050 + }, + { + "epoch": 84.04192139737991, + "grad_norm": 0.014379646629095078, + "learning_rate": 0.0006, + "loss": 4.0201239585876465, + "step": 6051 + }, + { + "epoch": 84.05589519650655, + "grad_norm": 0.014871201477944851, + "learning_rate": 0.0006, + "loss": 4.0820698738098145, + "step": 6052 + }, + { + "epoch": 84.06986899563319, + "grad_norm": 0.013085179962217808, + "learning_rate": 0.0006, + "loss": 4.223984718322754, + "step": 6053 + }, + { + "epoch": 84.08384279475983, + "grad_norm": 0.012485949322581291, + "learning_rate": 0.0006, + "loss": 4.182206153869629, + "step": 6054 + }, + { + "epoch": 84.09781659388646, + "grad_norm": 0.013653291389346123, + "learning_rate": 0.0006, + "loss": 4.142557144165039, + "step": 6055 + }, + { + "epoch": 84.1117903930131, + "grad_norm": 0.013930532149970531, + "learning_rate": 0.0006, + "loss": 4.13972282409668, + "step": 6056 + }, + { + "epoch": 84.12576419213974, + "grad_norm": 0.01297379657626152, + "learning_rate": 0.0006, + "loss": 4.144664764404297, + "step": 6057 + }, + { + "epoch": 84.13973799126637, + "grad_norm": 0.013210492208600044, + "learning_rate": 0.0006, + "loss": 4.253764629364014, + "step": 6058 + }, + { + "epoch": 84.15371179039302, + "grad_norm": 0.014650176279246807, + "learning_rate": 0.0006, + "loss": 4.246866226196289, + "step": 6059 + }, + { + "epoch": 84.16768558951965, + "grad_norm": 0.01482292078435421, + "learning_rate": 0.0006, + "loss": 4.140187740325928, + "step": 6060 + }, + { + "epoch": 84.18165938864628, + "grad_norm": 0.01592552848160267, + "learning_rate": 0.0006, + "loss": 4.170705318450928, + "step": 6061 + }, + { + "epoch": 84.19563318777293, + "grad_norm": 0.017152156680822372, + "learning_rate": 0.0006, + "loss": 4.126092433929443, + "step": 6062 + }, + { + "epoch": 84.20960698689956, + "grad_norm": 0.015880905091762543, + "learning_rate": 0.0006, + "loss": 4.161938190460205, + "step": 6063 + }, + { + "epoch": 84.2235807860262, + "grad_norm": 0.015451471321284771, + "learning_rate": 0.0006, + "loss": 4.263461112976074, + "step": 6064 + }, + { + "epoch": 84.23755458515284, + "grad_norm": 0.014365996234118938, + "learning_rate": 0.0006, + "loss": 4.14488410949707, + "step": 6065 + }, + { + "epoch": 84.25152838427948, + "grad_norm": 0.016364749521017075, + "learning_rate": 0.0006, + "loss": 4.112489700317383, + "step": 6066 + }, + { + "epoch": 84.26550218340611, + "grad_norm": 0.01791754551231861, + "learning_rate": 0.0006, + "loss": 4.133513450622559, + "step": 6067 + }, + { + "epoch": 84.27947598253274, + "grad_norm": 0.016609208658337593, + "learning_rate": 0.0006, + "loss": 4.201829433441162, + "step": 6068 + }, + { + "epoch": 84.29344978165939, + "grad_norm": 0.017587820068001747, + "learning_rate": 0.0006, + "loss": 4.203151702880859, + "step": 6069 + }, + { + "epoch": 84.30742358078602, + "grad_norm": 0.017047742381691933, + "learning_rate": 0.0006, + "loss": 4.077008247375488, + "step": 6070 + }, + { + "epoch": 84.32139737991267, + "grad_norm": 0.01524646207690239, + "learning_rate": 0.0006, + "loss": 4.198038101196289, + "step": 6071 + }, + { + "epoch": 84.3353711790393, + "grad_norm": 0.014196421951055527, + "learning_rate": 0.0006, + "loss": 4.164245128631592, + "step": 6072 + }, + { + "epoch": 84.34934497816595, + "grad_norm": 0.016291776672005653, + "learning_rate": 0.0006, + "loss": 4.150152206420898, + "step": 6073 + }, + { + "epoch": 84.36331877729258, + "grad_norm": 0.015627920627593994, + "learning_rate": 0.0006, + "loss": 4.110006332397461, + "step": 6074 + }, + { + "epoch": 84.37729257641921, + "grad_norm": 0.013251487165689468, + "learning_rate": 0.0006, + "loss": 4.101463317871094, + "step": 6075 + }, + { + "epoch": 84.39126637554585, + "grad_norm": 0.013601605780422688, + "learning_rate": 0.0006, + "loss": 4.191105842590332, + "step": 6076 + }, + { + "epoch": 84.40524017467249, + "grad_norm": 0.014878741465508938, + "learning_rate": 0.0006, + "loss": 4.153046131134033, + "step": 6077 + }, + { + "epoch": 84.41921397379913, + "grad_norm": 0.014783897437155247, + "learning_rate": 0.0006, + "loss": 4.362000465393066, + "step": 6078 + }, + { + "epoch": 84.43318777292576, + "grad_norm": 0.014221332035958767, + "learning_rate": 0.0006, + "loss": 4.144364833831787, + "step": 6079 + }, + { + "epoch": 84.44716157205241, + "grad_norm": 0.014911034144461155, + "learning_rate": 0.0006, + "loss": 4.088857650756836, + "step": 6080 + }, + { + "epoch": 84.46113537117904, + "grad_norm": 0.014924770221114159, + "learning_rate": 0.0006, + "loss": 4.1677446365356445, + "step": 6081 + }, + { + "epoch": 84.47510917030567, + "grad_norm": 0.014394409023225307, + "learning_rate": 0.0006, + "loss": 4.087103843688965, + "step": 6082 + }, + { + "epoch": 84.48908296943232, + "grad_norm": 0.013926676474511623, + "learning_rate": 0.0006, + "loss": 3.9972851276397705, + "step": 6083 + }, + { + "epoch": 84.50305676855895, + "grad_norm": 0.013069954700767994, + "learning_rate": 0.0006, + "loss": 3.998237133026123, + "step": 6084 + }, + { + "epoch": 84.5170305676856, + "grad_norm": 0.012748011387884617, + "learning_rate": 0.0006, + "loss": 4.195858955383301, + "step": 6085 + }, + { + "epoch": 84.53100436681223, + "grad_norm": 0.012278077192604542, + "learning_rate": 0.0006, + "loss": 4.102165222167969, + "step": 6086 + }, + { + "epoch": 84.54497816593886, + "grad_norm": 0.014259163290262222, + "learning_rate": 0.0006, + "loss": 4.092750549316406, + "step": 6087 + }, + { + "epoch": 84.5589519650655, + "grad_norm": 0.014553683809936047, + "learning_rate": 0.0006, + "loss": 4.1412763595581055, + "step": 6088 + }, + { + "epoch": 84.57292576419214, + "grad_norm": 0.013605243526399136, + "learning_rate": 0.0006, + "loss": 4.192173480987549, + "step": 6089 + }, + { + "epoch": 84.58689956331878, + "grad_norm": 0.014259358868002892, + "learning_rate": 0.0006, + "loss": 4.356374740600586, + "step": 6090 + }, + { + "epoch": 84.60087336244541, + "grad_norm": 0.014166025444865227, + "learning_rate": 0.0006, + "loss": 4.165982246398926, + "step": 6091 + }, + { + "epoch": 84.61484716157206, + "grad_norm": 0.012198803946375847, + "learning_rate": 0.0006, + "loss": 4.287566184997559, + "step": 6092 + }, + { + "epoch": 84.62882096069869, + "grad_norm": 0.013794238679111004, + "learning_rate": 0.0006, + "loss": 4.199882984161377, + "step": 6093 + }, + { + "epoch": 84.64279475982532, + "grad_norm": 0.014757642522454262, + "learning_rate": 0.0006, + "loss": 4.116573333740234, + "step": 6094 + }, + { + "epoch": 84.65676855895197, + "grad_norm": 0.014591689221560955, + "learning_rate": 0.0006, + "loss": 4.240896224975586, + "step": 6095 + }, + { + "epoch": 84.6707423580786, + "grad_norm": 0.015217579901218414, + "learning_rate": 0.0006, + "loss": 4.257142066955566, + "step": 6096 + }, + { + "epoch": 84.68471615720524, + "grad_norm": 0.015371371991932392, + "learning_rate": 0.0006, + "loss": 4.181037425994873, + "step": 6097 + }, + { + "epoch": 84.69868995633188, + "grad_norm": 0.014412983320653439, + "learning_rate": 0.0006, + "loss": 4.2207794189453125, + "step": 6098 + }, + { + "epoch": 84.71266375545852, + "grad_norm": 0.013773671351373196, + "learning_rate": 0.0006, + "loss": 4.211123466491699, + "step": 6099 + }, + { + "epoch": 84.72663755458515, + "grad_norm": 0.012838046066462994, + "learning_rate": 0.0006, + "loss": 4.175459384918213, + "step": 6100 + }, + { + "epoch": 84.74061135371178, + "grad_norm": 0.014094929210841656, + "learning_rate": 0.0006, + "loss": 4.1724958419799805, + "step": 6101 + }, + { + "epoch": 84.75458515283843, + "grad_norm": 0.014573262073099613, + "learning_rate": 0.0006, + "loss": 4.15885066986084, + "step": 6102 + }, + { + "epoch": 84.76855895196506, + "grad_norm": 0.011964819394052029, + "learning_rate": 0.0006, + "loss": 3.9804892539978027, + "step": 6103 + }, + { + "epoch": 84.78253275109171, + "grad_norm": 0.014140539802610874, + "learning_rate": 0.0006, + "loss": 4.203708648681641, + "step": 6104 + }, + { + "epoch": 84.79650655021834, + "grad_norm": 0.016618940979242325, + "learning_rate": 0.0006, + "loss": 4.147684097290039, + "step": 6105 + }, + { + "epoch": 84.81048034934499, + "grad_norm": 0.015094385482370853, + "learning_rate": 0.0006, + "loss": 4.392223834991455, + "step": 6106 + }, + { + "epoch": 84.82445414847162, + "grad_norm": 0.014386717230081558, + "learning_rate": 0.0006, + "loss": 4.253332138061523, + "step": 6107 + }, + { + "epoch": 84.83842794759825, + "grad_norm": 0.013877755962312222, + "learning_rate": 0.0006, + "loss": 4.075056076049805, + "step": 6108 + }, + { + "epoch": 84.8524017467249, + "grad_norm": 0.014062155969440937, + "learning_rate": 0.0006, + "loss": 4.072048664093018, + "step": 6109 + }, + { + "epoch": 84.86637554585153, + "grad_norm": 0.014445619657635689, + "learning_rate": 0.0006, + "loss": 4.151832580566406, + "step": 6110 + }, + { + "epoch": 84.88034934497817, + "grad_norm": 0.015116363763809204, + "learning_rate": 0.0006, + "loss": 4.088900566101074, + "step": 6111 + }, + { + "epoch": 84.8943231441048, + "grad_norm": 0.014425474219024181, + "learning_rate": 0.0006, + "loss": 4.200483322143555, + "step": 6112 + }, + { + "epoch": 84.90829694323143, + "grad_norm": 0.01449573878198862, + "learning_rate": 0.0006, + "loss": 4.239742279052734, + "step": 6113 + }, + { + "epoch": 84.92227074235808, + "grad_norm": 0.015698261559009552, + "learning_rate": 0.0006, + "loss": 4.197391510009766, + "step": 6114 + }, + { + "epoch": 84.93624454148471, + "grad_norm": 0.014545733109116554, + "learning_rate": 0.0006, + "loss": 4.035508155822754, + "step": 6115 + }, + { + "epoch": 84.95021834061136, + "grad_norm": 0.012796483002603054, + "learning_rate": 0.0006, + "loss": 4.195248126983643, + "step": 6116 + }, + { + "epoch": 84.96419213973799, + "grad_norm": 0.01264485064893961, + "learning_rate": 0.0006, + "loss": 4.158118724822998, + "step": 6117 + }, + { + "epoch": 84.97816593886463, + "grad_norm": 0.013695276342332363, + "learning_rate": 0.0006, + "loss": 4.255288124084473, + "step": 6118 + }, + { + "epoch": 84.99213973799127, + "grad_norm": 0.013459564186632633, + "learning_rate": 0.0006, + "loss": 4.204586505889893, + "step": 6119 + }, + { + "epoch": 85.0, + "grad_norm": 0.014312652871012688, + "learning_rate": 0.0006, + "loss": 4.167879104614258, + "step": 6120 + }, + { + "epoch": 85.0, + "eval_loss": 4.608152389526367, + "eval_runtime": 57.1903, + "eval_samples_per_second": 42.7, + "eval_steps_per_second": 1.346, + "step": 6120 + }, + { + "epoch": 85.01397379912663, + "grad_norm": 0.013918368145823479, + "learning_rate": 0.0006, + "loss": 4.129991054534912, + "step": 6121 + }, + { + "epoch": 85.02794759825328, + "grad_norm": 0.01604384370148182, + "learning_rate": 0.0006, + "loss": 4.233926773071289, + "step": 6122 + }, + { + "epoch": 85.04192139737991, + "grad_norm": 0.016565652564167976, + "learning_rate": 0.0006, + "loss": 4.129604339599609, + "step": 6123 + }, + { + "epoch": 85.05589519650655, + "grad_norm": 0.0149751678109169, + "learning_rate": 0.0006, + "loss": 4.165839195251465, + "step": 6124 + }, + { + "epoch": 85.06986899563319, + "grad_norm": 0.013812461867928505, + "learning_rate": 0.0006, + "loss": 4.153792858123779, + "step": 6125 + }, + { + "epoch": 85.08384279475983, + "grad_norm": 0.014432745985686779, + "learning_rate": 0.0006, + "loss": 4.170895576477051, + "step": 6126 + }, + { + "epoch": 85.09781659388646, + "grad_norm": 0.015208868309855461, + "learning_rate": 0.0006, + "loss": 4.10639762878418, + "step": 6127 + }, + { + "epoch": 85.1117903930131, + "grad_norm": 0.013209272176027298, + "learning_rate": 0.0006, + "loss": 4.200364112854004, + "step": 6128 + }, + { + "epoch": 85.12576419213974, + "grad_norm": 0.014807002618908882, + "learning_rate": 0.0006, + "loss": 4.172666549682617, + "step": 6129 + }, + { + "epoch": 85.13973799126637, + "grad_norm": 0.015662945806980133, + "learning_rate": 0.0006, + "loss": 4.145219326019287, + "step": 6130 + }, + { + "epoch": 85.15371179039302, + "grad_norm": 0.014598497189581394, + "learning_rate": 0.0006, + "loss": 4.275618553161621, + "step": 6131 + }, + { + "epoch": 85.16768558951965, + "grad_norm": 0.013778576627373695, + "learning_rate": 0.0006, + "loss": 4.204716682434082, + "step": 6132 + }, + { + "epoch": 85.18165938864628, + "grad_norm": 0.015762265771627426, + "learning_rate": 0.0006, + "loss": 4.145524024963379, + "step": 6133 + }, + { + "epoch": 85.19563318777293, + "grad_norm": 0.015579195693135262, + "learning_rate": 0.0006, + "loss": 4.150625228881836, + "step": 6134 + }, + { + "epoch": 85.20960698689956, + "grad_norm": 0.016641216352581978, + "learning_rate": 0.0006, + "loss": 4.141819953918457, + "step": 6135 + }, + { + "epoch": 85.2235807860262, + "grad_norm": 0.01632271707057953, + "learning_rate": 0.0006, + "loss": 4.030620574951172, + "step": 6136 + }, + { + "epoch": 85.23755458515284, + "grad_norm": 0.015564918518066406, + "learning_rate": 0.0006, + "loss": 4.147753715515137, + "step": 6137 + }, + { + "epoch": 85.25152838427948, + "grad_norm": 0.0161129217594862, + "learning_rate": 0.0006, + "loss": 4.128524303436279, + "step": 6138 + }, + { + "epoch": 85.26550218340611, + "grad_norm": 0.017766550183296204, + "learning_rate": 0.0006, + "loss": 4.134856224060059, + "step": 6139 + }, + { + "epoch": 85.27947598253274, + "grad_norm": 0.01869054324924946, + "learning_rate": 0.0006, + "loss": 4.231565475463867, + "step": 6140 + }, + { + "epoch": 85.29344978165939, + "grad_norm": 0.01917010359466076, + "learning_rate": 0.0006, + "loss": 4.21175479888916, + "step": 6141 + }, + { + "epoch": 85.30742358078602, + "grad_norm": 0.018116485327482224, + "learning_rate": 0.0006, + "loss": 4.122644424438477, + "step": 6142 + }, + { + "epoch": 85.32139737991267, + "grad_norm": 0.015949107706546783, + "learning_rate": 0.0006, + "loss": 4.135458946228027, + "step": 6143 + }, + { + "epoch": 85.3353711790393, + "grad_norm": 0.014783683232963085, + "learning_rate": 0.0006, + "loss": 4.118994235992432, + "step": 6144 + }, + { + "epoch": 85.34934497816595, + "grad_norm": 0.015406692400574684, + "learning_rate": 0.0006, + "loss": 4.157268524169922, + "step": 6145 + }, + { + "epoch": 85.36331877729258, + "grad_norm": 0.013893205672502518, + "learning_rate": 0.0006, + "loss": 4.056166648864746, + "step": 6146 + }, + { + "epoch": 85.37729257641921, + "grad_norm": 0.014380923472344875, + "learning_rate": 0.0006, + "loss": 4.206905364990234, + "step": 6147 + }, + { + "epoch": 85.39126637554585, + "grad_norm": 0.015669258311390877, + "learning_rate": 0.0006, + "loss": 4.213881492614746, + "step": 6148 + }, + { + "epoch": 85.40524017467249, + "grad_norm": 0.01553428452461958, + "learning_rate": 0.0006, + "loss": 4.1814374923706055, + "step": 6149 + }, + { + "epoch": 85.41921397379913, + "grad_norm": 0.01697700284421444, + "learning_rate": 0.0006, + "loss": 4.083939552307129, + "step": 6150 + }, + { + "epoch": 85.43318777292576, + "grad_norm": 0.016639070585370064, + "learning_rate": 0.0006, + "loss": 4.279311180114746, + "step": 6151 + }, + { + "epoch": 85.44716157205241, + "grad_norm": 0.016285281628370285, + "learning_rate": 0.0006, + "loss": 4.244845390319824, + "step": 6152 + }, + { + "epoch": 85.46113537117904, + "grad_norm": 0.016071317717432976, + "learning_rate": 0.0006, + "loss": 4.089814186096191, + "step": 6153 + }, + { + "epoch": 85.47510917030567, + "grad_norm": 0.01640418916940689, + "learning_rate": 0.0006, + "loss": 4.168088912963867, + "step": 6154 + }, + { + "epoch": 85.48908296943232, + "grad_norm": 0.014781414531171322, + "learning_rate": 0.0006, + "loss": 4.196092128753662, + "step": 6155 + }, + { + "epoch": 85.50305676855895, + "grad_norm": 0.013393213972449303, + "learning_rate": 0.0006, + "loss": 4.193429946899414, + "step": 6156 + }, + { + "epoch": 85.5170305676856, + "grad_norm": 0.01467406190931797, + "learning_rate": 0.0006, + "loss": 4.216714859008789, + "step": 6157 + }, + { + "epoch": 85.53100436681223, + "grad_norm": 0.014950765296816826, + "learning_rate": 0.0006, + "loss": 4.093854904174805, + "step": 6158 + }, + { + "epoch": 85.54497816593886, + "grad_norm": 0.014655301347374916, + "learning_rate": 0.0006, + "loss": 4.09483528137207, + "step": 6159 + }, + { + "epoch": 85.5589519650655, + "grad_norm": 0.01425520796328783, + "learning_rate": 0.0006, + "loss": 4.090396881103516, + "step": 6160 + }, + { + "epoch": 85.57292576419214, + "grad_norm": 0.01403278298676014, + "learning_rate": 0.0006, + "loss": 4.088724136352539, + "step": 6161 + }, + { + "epoch": 85.58689956331878, + "grad_norm": 0.01364967692643404, + "learning_rate": 0.0006, + "loss": 4.095558166503906, + "step": 6162 + }, + { + "epoch": 85.60087336244541, + "grad_norm": 0.013085267506539822, + "learning_rate": 0.0006, + "loss": 4.163962364196777, + "step": 6163 + }, + { + "epoch": 85.61484716157206, + "grad_norm": 0.0136906448751688, + "learning_rate": 0.0006, + "loss": 4.197134017944336, + "step": 6164 + }, + { + "epoch": 85.62882096069869, + "grad_norm": 0.013472451828420162, + "learning_rate": 0.0006, + "loss": 4.1578369140625, + "step": 6165 + }, + { + "epoch": 85.64279475982532, + "grad_norm": 0.012329773046076298, + "learning_rate": 0.0006, + "loss": 4.231907844543457, + "step": 6166 + }, + { + "epoch": 85.65676855895197, + "grad_norm": 0.011977200396358967, + "learning_rate": 0.0006, + "loss": 4.1552958488464355, + "step": 6167 + }, + { + "epoch": 85.6707423580786, + "grad_norm": 0.013049956411123276, + "learning_rate": 0.0006, + "loss": 4.17852783203125, + "step": 6168 + }, + { + "epoch": 85.68471615720524, + "grad_norm": 0.011440741829574108, + "learning_rate": 0.0006, + "loss": 4.144646644592285, + "step": 6169 + }, + { + "epoch": 85.69868995633188, + "grad_norm": 0.012472936883568764, + "learning_rate": 0.0006, + "loss": 4.2306413650512695, + "step": 6170 + }, + { + "epoch": 85.71266375545852, + "grad_norm": 0.013015356846153736, + "learning_rate": 0.0006, + "loss": 4.193439960479736, + "step": 6171 + }, + { + "epoch": 85.72663755458515, + "grad_norm": 0.013832319527864456, + "learning_rate": 0.0006, + "loss": 4.18073034286499, + "step": 6172 + }, + { + "epoch": 85.74061135371178, + "grad_norm": 0.014613982290029526, + "learning_rate": 0.0006, + "loss": 4.223531246185303, + "step": 6173 + }, + { + "epoch": 85.75458515283843, + "grad_norm": 0.014209083281457424, + "learning_rate": 0.0006, + "loss": 4.142666816711426, + "step": 6174 + }, + { + "epoch": 85.76855895196506, + "grad_norm": 0.013396669179201126, + "learning_rate": 0.0006, + "loss": 4.185786247253418, + "step": 6175 + }, + { + "epoch": 85.78253275109171, + "grad_norm": 0.01364929135888815, + "learning_rate": 0.0006, + "loss": 4.125226974487305, + "step": 6176 + }, + { + "epoch": 85.79650655021834, + "grad_norm": 0.012561003677546978, + "learning_rate": 0.0006, + "loss": 4.084519386291504, + "step": 6177 + }, + { + "epoch": 85.81048034934499, + "grad_norm": 0.013508481904864311, + "learning_rate": 0.0006, + "loss": 4.193173885345459, + "step": 6178 + }, + { + "epoch": 85.82445414847162, + "grad_norm": 0.014154613018035889, + "learning_rate": 0.0006, + "loss": 4.257272720336914, + "step": 6179 + }, + { + "epoch": 85.83842794759825, + "grad_norm": 0.012478531338274479, + "learning_rate": 0.0006, + "loss": 4.161380767822266, + "step": 6180 + }, + { + "epoch": 85.8524017467249, + "grad_norm": 0.013468451797962189, + "learning_rate": 0.0006, + "loss": 4.121395111083984, + "step": 6181 + }, + { + "epoch": 85.86637554585153, + "grad_norm": 0.015155194327235222, + "learning_rate": 0.0006, + "loss": 4.242815971374512, + "step": 6182 + }, + { + "epoch": 85.88034934497817, + "grad_norm": 0.016223272308707237, + "learning_rate": 0.0006, + "loss": 4.176027297973633, + "step": 6183 + }, + { + "epoch": 85.8943231441048, + "grad_norm": 0.013741295784711838, + "learning_rate": 0.0006, + "loss": 4.131257057189941, + "step": 6184 + }, + { + "epoch": 85.90829694323143, + "grad_norm": 0.014982067979872227, + "learning_rate": 0.0006, + "loss": 4.237896919250488, + "step": 6185 + }, + { + "epoch": 85.92227074235808, + "grad_norm": 0.018197322264313698, + "learning_rate": 0.0006, + "loss": 4.154962539672852, + "step": 6186 + }, + { + "epoch": 85.93624454148471, + "grad_norm": 0.01971200481057167, + "learning_rate": 0.0006, + "loss": 4.16819429397583, + "step": 6187 + }, + { + "epoch": 85.95021834061136, + "grad_norm": 0.018008720129728317, + "learning_rate": 0.0006, + "loss": 4.146164417266846, + "step": 6188 + }, + { + "epoch": 85.96419213973799, + "grad_norm": 0.015859264880418777, + "learning_rate": 0.0006, + "loss": 4.208099365234375, + "step": 6189 + }, + { + "epoch": 85.97816593886463, + "grad_norm": 0.013846252113580704, + "learning_rate": 0.0006, + "loss": 4.1881184577941895, + "step": 6190 + }, + { + "epoch": 85.99213973799127, + "grad_norm": 0.018641939386725426, + "learning_rate": 0.0006, + "loss": 4.127632141113281, + "step": 6191 + }, + { + "epoch": 86.0, + "grad_norm": 0.02385227382183075, + "learning_rate": 0.0006, + "loss": 4.281900405883789, + "step": 6192 + }, + { + "epoch": 86.0, + "eval_loss": 4.625458717346191, + "eval_runtime": 57.0634, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6192 + }, + { + "epoch": 86.01397379912663, + "grad_norm": 0.019032124429941177, + "learning_rate": 0.0006, + "loss": 4.2163190841674805, + "step": 6193 + }, + { + "epoch": 86.02794759825328, + "grad_norm": 0.018049925565719604, + "learning_rate": 0.0006, + "loss": 4.1534037590026855, + "step": 6194 + }, + { + "epoch": 86.04192139737991, + "grad_norm": 0.019116360694169998, + "learning_rate": 0.0006, + "loss": 4.212411403656006, + "step": 6195 + }, + { + "epoch": 86.05589519650655, + "grad_norm": 0.018967336043715477, + "learning_rate": 0.0006, + "loss": 4.190918922424316, + "step": 6196 + }, + { + "epoch": 86.06986899563319, + "grad_norm": 0.01945611648261547, + "learning_rate": 0.0006, + "loss": 4.18459415435791, + "step": 6197 + }, + { + "epoch": 86.08384279475983, + "grad_norm": 0.018104076385498047, + "learning_rate": 0.0006, + "loss": 4.0942816734313965, + "step": 6198 + }, + { + "epoch": 86.09781659388646, + "grad_norm": 0.01815992221236229, + "learning_rate": 0.0006, + "loss": 4.011547565460205, + "step": 6199 + }, + { + "epoch": 86.1117903930131, + "grad_norm": 0.017574431374669075, + "learning_rate": 0.0006, + "loss": 4.051593780517578, + "step": 6200 + }, + { + "epoch": 86.12576419213974, + "grad_norm": 0.01516508124768734, + "learning_rate": 0.0006, + "loss": 4.239673614501953, + "step": 6201 + }, + { + "epoch": 86.13973799126637, + "grad_norm": 0.015689702704548836, + "learning_rate": 0.0006, + "loss": 4.1085734367370605, + "step": 6202 + }, + { + "epoch": 86.15371179039302, + "grad_norm": 0.01596900261938572, + "learning_rate": 0.0006, + "loss": 4.1509599685668945, + "step": 6203 + }, + { + "epoch": 86.16768558951965, + "grad_norm": 0.01550096832215786, + "learning_rate": 0.0006, + "loss": 4.160238265991211, + "step": 6204 + }, + { + "epoch": 86.18165938864628, + "grad_norm": 0.014601529575884342, + "learning_rate": 0.0006, + "loss": 4.029219150543213, + "step": 6205 + }, + { + "epoch": 86.19563318777293, + "grad_norm": 0.014006740413606167, + "learning_rate": 0.0006, + "loss": 4.117501258850098, + "step": 6206 + }, + { + "epoch": 86.20960698689956, + "grad_norm": 0.013191957026720047, + "learning_rate": 0.0006, + "loss": 4.252002716064453, + "step": 6207 + }, + { + "epoch": 86.2235807860262, + "grad_norm": 0.013031814247369766, + "learning_rate": 0.0006, + "loss": 4.209517478942871, + "step": 6208 + }, + { + "epoch": 86.23755458515284, + "grad_norm": 0.013605714775621891, + "learning_rate": 0.0006, + "loss": 4.176309585571289, + "step": 6209 + }, + { + "epoch": 86.25152838427948, + "grad_norm": 0.012753087095916271, + "learning_rate": 0.0006, + "loss": 4.097990036010742, + "step": 6210 + }, + { + "epoch": 86.26550218340611, + "grad_norm": 0.012940628454089165, + "learning_rate": 0.0006, + "loss": 4.085589408874512, + "step": 6211 + }, + { + "epoch": 86.27947598253274, + "grad_norm": 0.014286619611084461, + "learning_rate": 0.0006, + "loss": 4.199261665344238, + "step": 6212 + }, + { + "epoch": 86.29344978165939, + "grad_norm": 0.014643525704741478, + "learning_rate": 0.0006, + "loss": 4.23718786239624, + "step": 6213 + }, + { + "epoch": 86.30742358078602, + "grad_norm": 0.015182147733867168, + "learning_rate": 0.0006, + "loss": 4.219855308532715, + "step": 6214 + }, + { + "epoch": 86.32139737991267, + "grad_norm": 0.01530102826654911, + "learning_rate": 0.0006, + "loss": 4.190094947814941, + "step": 6215 + }, + { + "epoch": 86.3353711790393, + "grad_norm": 0.015492603182792664, + "learning_rate": 0.0006, + "loss": 4.153780460357666, + "step": 6216 + }, + { + "epoch": 86.34934497816595, + "grad_norm": 0.015339907258749008, + "learning_rate": 0.0006, + "loss": 4.182865619659424, + "step": 6217 + }, + { + "epoch": 86.36331877729258, + "grad_norm": 0.015576880425214767, + "learning_rate": 0.0006, + "loss": 4.196417331695557, + "step": 6218 + }, + { + "epoch": 86.37729257641921, + "grad_norm": 0.014805737882852554, + "learning_rate": 0.0006, + "loss": 4.265018463134766, + "step": 6219 + }, + { + "epoch": 86.39126637554585, + "grad_norm": 0.01597665622830391, + "learning_rate": 0.0006, + "loss": 4.2864837646484375, + "step": 6220 + }, + { + "epoch": 86.40524017467249, + "grad_norm": 0.01693643443286419, + "learning_rate": 0.0006, + "loss": 4.118192672729492, + "step": 6221 + }, + { + "epoch": 86.41921397379913, + "grad_norm": 0.016648951917886734, + "learning_rate": 0.0006, + "loss": 4.23712682723999, + "step": 6222 + }, + { + "epoch": 86.43318777292576, + "grad_norm": 0.014809766784310341, + "learning_rate": 0.0006, + "loss": 4.251391410827637, + "step": 6223 + }, + { + "epoch": 86.44716157205241, + "grad_norm": 0.014333373866975307, + "learning_rate": 0.0006, + "loss": 4.17243766784668, + "step": 6224 + }, + { + "epoch": 86.46113537117904, + "grad_norm": 0.013912711292505264, + "learning_rate": 0.0006, + "loss": 4.250836372375488, + "step": 6225 + }, + { + "epoch": 86.47510917030567, + "grad_norm": 0.014821341261267662, + "learning_rate": 0.0006, + "loss": 4.073995590209961, + "step": 6226 + }, + { + "epoch": 86.48908296943232, + "grad_norm": 0.016634922474622726, + "learning_rate": 0.0006, + "loss": 4.155867576599121, + "step": 6227 + }, + { + "epoch": 86.50305676855895, + "grad_norm": 0.01748695969581604, + "learning_rate": 0.0006, + "loss": 4.139344692230225, + "step": 6228 + }, + { + "epoch": 86.5170305676856, + "grad_norm": 0.01569363661110401, + "learning_rate": 0.0006, + "loss": 4.23923397064209, + "step": 6229 + }, + { + "epoch": 86.53100436681223, + "grad_norm": 0.014037121087312698, + "learning_rate": 0.0006, + "loss": 4.200334072113037, + "step": 6230 + }, + { + "epoch": 86.54497816593886, + "grad_norm": 0.01401279866695404, + "learning_rate": 0.0006, + "loss": 4.035246849060059, + "step": 6231 + }, + { + "epoch": 86.5589519650655, + "grad_norm": 0.014603731222450733, + "learning_rate": 0.0006, + "loss": 4.141385555267334, + "step": 6232 + }, + { + "epoch": 86.57292576419214, + "grad_norm": 0.015113704837858677, + "learning_rate": 0.0006, + "loss": 4.130824089050293, + "step": 6233 + }, + { + "epoch": 86.58689956331878, + "grad_norm": 0.01588696800172329, + "learning_rate": 0.0006, + "loss": 4.05495023727417, + "step": 6234 + }, + { + "epoch": 86.60087336244541, + "grad_norm": 0.01593640074133873, + "learning_rate": 0.0006, + "loss": 4.22935152053833, + "step": 6235 + }, + { + "epoch": 86.61484716157206, + "grad_norm": 0.014292445965111256, + "learning_rate": 0.0006, + "loss": 4.309839248657227, + "step": 6236 + }, + { + "epoch": 86.62882096069869, + "grad_norm": 0.015471610240638256, + "learning_rate": 0.0006, + "loss": 4.268278121948242, + "step": 6237 + }, + { + "epoch": 86.64279475982532, + "grad_norm": 0.016643544659018517, + "learning_rate": 0.0006, + "loss": 4.135109901428223, + "step": 6238 + }, + { + "epoch": 86.65676855895197, + "grad_norm": 0.014733029529452324, + "learning_rate": 0.0006, + "loss": 4.060197830200195, + "step": 6239 + }, + { + "epoch": 86.6707423580786, + "grad_norm": 0.014869216829538345, + "learning_rate": 0.0006, + "loss": 4.126917362213135, + "step": 6240 + }, + { + "epoch": 86.68471615720524, + "grad_norm": 0.01608203910291195, + "learning_rate": 0.0006, + "loss": 4.183716773986816, + "step": 6241 + }, + { + "epoch": 86.69868995633188, + "grad_norm": 0.016178175806999207, + "learning_rate": 0.0006, + "loss": 4.118875503540039, + "step": 6242 + }, + { + "epoch": 86.71266375545852, + "grad_norm": 0.014576597139239311, + "learning_rate": 0.0006, + "loss": 4.261541366577148, + "step": 6243 + }, + { + "epoch": 86.72663755458515, + "grad_norm": 0.014466420747339725, + "learning_rate": 0.0006, + "loss": 4.307284355163574, + "step": 6244 + }, + { + "epoch": 86.74061135371178, + "grad_norm": 0.014042644761502743, + "learning_rate": 0.0006, + "loss": 4.2729997634887695, + "step": 6245 + }, + { + "epoch": 86.75458515283843, + "grad_norm": 0.014589524827897549, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6246 + }, + { + "epoch": 86.76855895196506, + "grad_norm": 0.018765483051538467, + "learning_rate": 0.0006, + "loss": 4.264769554138184, + "step": 6247 + }, + { + "epoch": 86.78253275109171, + "grad_norm": 0.019397731870412827, + "learning_rate": 0.0006, + "loss": 4.183379173278809, + "step": 6248 + }, + { + "epoch": 86.79650655021834, + "grad_norm": 0.017525402829051018, + "learning_rate": 0.0006, + "loss": 4.1494245529174805, + "step": 6249 + }, + { + "epoch": 86.81048034934499, + "grad_norm": 0.016310200095176697, + "learning_rate": 0.0006, + "loss": 4.149104118347168, + "step": 6250 + }, + { + "epoch": 86.82445414847162, + "grad_norm": 0.01506423857063055, + "learning_rate": 0.0006, + "loss": 4.22091817855835, + "step": 6251 + }, + { + "epoch": 86.83842794759825, + "grad_norm": 0.014200284145772457, + "learning_rate": 0.0006, + "loss": 4.177569389343262, + "step": 6252 + }, + { + "epoch": 86.8524017467249, + "grad_norm": 0.015016552060842514, + "learning_rate": 0.0006, + "loss": 4.089859962463379, + "step": 6253 + }, + { + "epoch": 86.86637554585153, + "grad_norm": 0.016054971143603325, + "learning_rate": 0.0006, + "loss": 4.157166004180908, + "step": 6254 + }, + { + "epoch": 86.88034934497817, + "grad_norm": 0.014388521201908588, + "learning_rate": 0.0006, + "loss": 4.1437578201293945, + "step": 6255 + }, + { + "epoch": 86.8943231441048, + "grad_norm": 0.015705369412899017, + "learning_rate": 0.0006, + "loss": 4.045689582824707, + "step": 6256 + }, + { + "epoch": 86.90829694323143, + "grad_norm": 0.014192395843565464, + "learning_rate": 0.0006, + "loss": 4.098793029785156, + "step": 6257 + }, + { + "epoch": 86.92227074235808, + "grad_norm": 0.013751034624874592, + "learning_rate": 0.0006, + "loss": 3.989920139312744, + "step": 6258 + }, + { + "epoch": 86.93624454148471, + "grad_norm": 0.01582193188369274, + "learning_rate": 0.0006, + "loss": 4.283449172973633, + "step": 6259 + }, + { + "epoch": 86.95021834061136, + "grad_norm": 0.01476013008505106, + "learning_rate": 0.0006, + "loss": 4.221401691436768, + "step": 6260 + }, + { + "epoch": 86.96419213973799, + "grad_norm": 0.01346337515860796, + "learning_rate": 0.0006, + "loss": 4.05896520614624, + "step": 6261 + }, + { + "epoch": 86.97816593886463, + "grad_norm": 0.012899307534098625, + "learning_rate": 0.0006, + "loss": 4.159560203552246, + "step": 6262 + }, + { + "epoch": 86.99213973799127, + "grad_norm": 0.013150574639439583, + "learning_rate": 0.0006, + "loss": 4.217330455780029, + "step": 6263 + }, + { + "epoch": 87.0, + "grad_norm": 0.014940370805561543, + "learning_rate": 0.0006, + "loss": 4.1611833572387695, + "step": 6264 + }, + { + "epoch": 87.0, + "eval_loss": 4.666043758392334, + "eval_runtime": 56.5978, + "eval_samples_per_second": 43.147, + "eval_steps_per_second": 1.36, + "step": 6264 + }, + { + "epoch": 87.01397379912663, + "grad_norm": 0.013721180148422718, + "learning_rate": 0.0006, + "loss": 4.142454147338867, + "step": 6265 + }, + { + "epoch": 87.02794759825328, + "grad_norm": 0.014805255457758904, + "learning_rate": 0.0006, + "loss": 4.121677875518799, + "step": 6266 + }, + { + "epoch": 87.04192139737991, + "grad_norm": 0.015576514415442944, + "learning_rate": 0.0006, + "loss": 4.1188435554504395, + "step": 6267 + }, + { + "epoch": 87.05589519650655, + "grad_norm": 0.01605502888560295, + "learning_rate": 0.0006, + "loss": 4.255428314208984, + "step": 6268 + }, + { + "epoch": 87.06986899563319, + "grad_norm": 0.014933343045413494, + "learning_rate": 0.0006, + "loss": 4.122596740722656, + "step": 6269 + }, + { + "epoch": 87.08384279475983, + "grad_norm": 0.014630894176661968, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 6270 + }, + { + "epoch": 87.09781659388646, + "grad_norm": 0.014211163856089115, + "learning_rate": 0.0006, + "loss": 4.058030128479004, + "step": 6271 + }, + { + "epoch": 87.1117903930131, + "grad_norm": 0.014060168527066708, + "learning_rate": 0.0006, + "loss": 4.146586894989014, + "step": 6272 + }, + { + "epoch": 87.12576419213974, + "grad_norm": 0.014740100130438805, + "learning_rate": 0.0006, + "loss": 4.077739715576172, + "step": 6273 + }, + { + "epoch": 87.13973799126637, + "grad_norm": 0.013970664702355862, + "learning_rate": 0.0006, + "loss": 4.038252830505371, + "step": 6274 + }, + { + "epoch": 87.15371179039302, + "grad_norm": 0.014399951323866844, + "learning_rate": 0.0006, + "loss": 4.143700122833252, + "step": 6275 + }, + { + "epoch": 87.16768558951965, + "grad_norm": 0.01496410183608532, + "learning_rate": 0.0006, + "loss": 4.105609893798828, + "step": 6276 + }, + { + "epoch": 87.18165938864628, + "grad_norm": 0.014287355355918407, + "learning_rate": 0.0006, + "loss": 4.173645973205566, + "step": 6277 + }, + { + "epoch": 87.19563318777293, + "grad_norm": 0.0141292130574584, + "learning_rate": 0.0006, + "loss": 4.068511486053467, + "step": 6278 + }, + { + "epoch": 87.20960698689956, + "grad_norm": 0.01393523346632719, + "learning_rate": 0.0006, + "loss": 4.060687065124512, + "step": 6279 + }, + { + "epoch": 87.2235807860262, + "grad_norm": 0.01431021187454462, + "learning_rate": 0.0006, + "loss": 4.219209671020508, + "step": 6280 + }, + { + "epoch": 87.23755458515284, + "grad_norm": 0.014856027439236641, + "learning_rate": 0.0006, + "loss": 4.146749973297119, + "step": 6281 + }, + { + "epoch": 87.25152838427948, + "grad_norm": 0.013677539303898811, + "learning_rate": 0.0006, + "loss": 3.9859797954559326, + "step": 6282 + }, + { + "epoch": 87.26550218340611, + "grad_norm": 0.014078967273235321, + "learning_rate": 0.0006, + "loss": 4.129668235778809, + "step": 6283 + }, + { + "epoch": 87.27947598253274, + "grad_norm": 0.015564394183456898, + "learning_rate": 0.0006, + "loss": 4.0007829666137695, + "step": 6284 + }, + { + "epoch": 87.29344978165939, + "grad_norm": 0.015077749267220497, + "learning_rate": 0.0006, + "loss": 4.112264156341553, + "step": 6285 + }, + { + "epoch": 87.30742358078602, + "grad_norm": 0.015341218560934067, + "learning_rate": 0.0006, + "loss": 4.112249374389648, + "step": 6286 + }, + { + "epoch": 87.32139737991267, + "grad_norm": 0.016399366781115532, + "learning_rate": 0.0006, + "loss": 4.176196098327637, + "step": 6287 + }, + { + "epoch": 87.3353711790393, + "grad_norm": 0.01599763333797455, + "learning_rate": 0.0006, + "loss": 4.163441181182861, + "step": 6288 + }, + { + "epoch": 87.34934497816595, + "grad_norm": 0.015439776703715324, + "learning_rate": 0.0006, + "loss": 4.11902379989624, + "step": 6289 + }, + { + "epoch": 87.36331877729258, + "grad_norm": 0.015909943729639053, + "learning_rate": 0.0006, + "loss": 4.16621732711792, + "step": 6290 + }, + { + "epoch": 87.37729257641921, + "grad_norm": 0.016710078343749046, + "learning_rate": 0.0006, + "loss": 4.044960975646973, + "step": 6291 + }, + { + "epoch": 87.39126637554585, + "grad_norm": 0.015883933752775192, + "learning_rate": 0.0006, + "loss": 4.065184593200684, + "step": 6292 + }, + { + "epoch": 87.40524017467249, + "grad_norm": 0.013912747614085674, + "learning_rate": 0.0006, + "loss": 4.102450370788574, + "step": 6293 + }, + { + "epoch": 87.41921397379913, + "grad_norm": 0.013170420192182064, + "learning_rate": 0.0006, + "loss": 4.234364032745361, + "step": 6294 + }, + { + "epoch": 87.43318777292576, + "grad_norm": 0.014043302275240421, + "learning_rate": 0.0006, + "loss": 4.129565238952637, + "step": 6295 + }, + { + "epoch": 87.44716157205241, + "grad_norm": 0.015010139904916286, + "learning_rate": 0.0006, + "loss": 4.166678428649902, + "step": 6296 + }, + { + "epoch": 87.46113537117904, + "grad_norm": 0.014721784740686417, + "learning_rate": 0.0006, + "loss": 4.182618618011475, + "step": 6297 + }, + { + "epoch": 87.47510917030567, + "grad_norm": 0.015713468194007874, + "learning_rate": 0.0006, + "loss": 4.103855133056641, + "step": 6298 + }, + { + "epoch": 87.48908296943232, + "grad_norm": 0.017204247415065765, + "learning_rate": 0.0006, + "loss": 4.140937328338623, + "step": 6299 + }, + { + "epoch": 87.50305676855895, + "grad_norm": 0.016856668516993523, + "learning_rate": 0.0006, + "loss": 4.147859573364258, + "step": 6300 + }, + { + "epoch": 87.5170305676856, + "grad_norm": 0.017205018550157547, + "learning_rate": 0.0006, + "loss": 4.1667327880859375, + "step": 6301 + }, + { + "epoch": 87.53100436681223, + "grad_norm": 0.014821539632976055, + "learning_rate": 0.0006, + "loss": 4.089935302734375, + "step": 6302 + }, + { + "epoch": 87.54497816593886, + "grad_norm": 0.01446484588086605, + "learning_rate": 0.0006, + "loss": 4.243657112121582, + "step": 6303 + }, + { + "epoch": 87.5589519650655, + "grad_norm": 0.014374484308063984, + "learning_rate": 0.0006, + "loss": 4.14356803894043, + "step": 6304 + }, + { + "epoch": 87.57292576419214, + "grad_norm": 0.01420109998434782, + "learning_rate": 0.0006, + "loss": 4.1273651123046875, + "step": 6305 + }, + { + "epoch": 87.58689956331878, + "grad_norm": 0.014810388907790184, + "learning_rate": 0.0006, + "loss": 4.222290515899658, + "step": 6306 + }, + { + "epoch": 87.60087336244541, + "grad_norm": 0.015826983377337456, + "learning_rate": 0.0006, + "loss": 4.137996196746826, + "step": 6307 + }, + { + "epoch": 87.61484716157206, + "grad_norm": 0.014457454904913902, + "learning_rate": 0.0006, + "loss": 4.0780744552612305, + "step": 6308 + }, + { + "epoch": 87.62882096069869, + "grad_norm": 0.014458604156970978, + "learning_rate": 0.0006, + "loss": 4.1287922859191895, + "step": 6309 + }, + { + "epoch": 87.64279475982532, + "grad_norm": 0.015282071195542812, + "learning_rate": 0.0006, + "loss": 4.189508438110352, + "step": 6310 + }, + { + "epoch": 87.65676855895197, + "grad_norm": 0.017536642029881477, + "learning_rate": 0.0006, + "loss": 4.070173263549805, + "step": 6311 + }, + { + "epoch": 87.6707423580786, + "grad_norm": 0.017303580418229103, + "learning_rate": 0.0006, + "loss": 4.119133949279785, + "step": 6312 + }, + { + "epoch": 87.68471615720524, + "grad_norm": 0.016148580238223076, + "learning_rate": 0.0006, + "loss": 3.994354486465454, + "step": 6313 + }, + { + "epoch": 87.69868995633188, + "grad_norm": 0.01706942543387413, + "learning_rate": 0.0006, + "loss": 4.183113098144531, + "step": 6314 + }, + { + "epoch": 87.71266375545852, + "grad_norm": 0.016971010714769363, + "learning_rate": 0.0006, + "loss": 4.104086875915527, + "step": 6315 + }, + { + "epoch": 87.72663755458515, + "grad_norm": 0.015516847372055054, + "learning_rate": 0.0006, + "loss": 4.2409772872924805, + "step": 6316 + }, + { + "epoch": 87.74061135371178, + "grad_norm": 0.015352269634604454, + "learning_rate": 0.0006, + "loss": 4.026427268981934, + "step": 6317 + }, + { + "epoch": 87.75458515283843, + "grad_norm": 0.017702415585517883, + "learning_rate": 0.0006, + "loss": 4.177840232849121, + "step": 6318 + }, + { + "epoch": 87.76855895196506, + "grad_norm": 0.01629541628062725, + "learning_rate": 0.0006, + "loss": 4.229039192199707, + "step": 6319 + }, + { + "epoch": 87.78253275109171, + "grad_norm": 0.01503711100667715, + "learning_rate": 0.0006, + "loss": 4.214350700378418, + "step": 6320 + }, + { + "epoch": 87.79650655021834, + "grad_norm": 0.014913883060216904, + "learning_rate": 0.0006, + "loss": 4.168880939483643, + "step": 6321 + }, + { + "epoch": 87.81048034934499, + "grad_norm": 0.014513909816741943, + "learning_rate": 0.0006, + "loss": 4.162267684936523, + "step": 6322 + }, + { + "epoch": 87.82445414847162, + "grad_norm": 0.014635873027145863, + "learning_rate": 0.0006, + "loss": 4.220067977905273, + "step": 6323 + }, + { + "epoch": 87.83842794759825, + "grad_norm": 0.017102796584367752, + "learning_rate": 0.0006, + "loss": 4.1493377685546875, + "step": 6324 + }, + { + "epoch": 87.8524017467249, + "grad_norm": 0.015533884055912495, + "learning_rate": 0.0006, + "loss": 4.0646209716796875, + "step": 6325 + }, + { + "epoch": 87.86637554585153, + "grad_norm": 0.014617550186812878, + "learning_rate": 0.0006, + "loss": 4.1390533447265625, + "step": 6326 + }, + { + "epoch": 87.88034934497817, + "grad_norm": 0.014142967760562897, + "learning_rate": 0.0006, + "loss": 4.133934020996094, + "step": 6327 + }, + { + "epoch": 87.8943231441048, + "grad_norm": 0.013182433322072029, + "learning_rate": 0.0006, + "loss": 4.125327110290527, + "step": 6328 + }, + { + "epoch": 87.90829694323143, + "grad_norm": 0.013928229920566082, + "learning_rate": 0.0006, + "loss": 4.142777442932129, + "step": 6329 + }, + { + "epoch": 87.92227074235808, + "grad_norm": 0.015337013639509678, + "learning_rate": 0.0006, + "loss": 4.140210151672363, + "step": 6330 + }, + { + "epoch": 87.93624454148471, + "grad_norm": 0.014873536303639412, + "learning_rate": 0.0006, + "loss": 4.102636337280273, + "step": 6331 + }, + { + "epoch": 87.95021834061136, + "grad_norm": 0.01387431938201189, + "learning_rate": 0.0006, + "loss": 4.149059295654297, + "step": 6332 + }, + { + "epoch": 87.96419213973799, + "grad_norm": 0.013234943151473999, + "learning_rate": 0.0006, + "loss": 4.1564459800720215, + "step": 6333 + }, + { + "epoch": 87.97816593886463, + "grad_norm": 0.013781017623841763, + "learning_rate": 0.0006, + "loss": 4.189839839935303, + "step": 6334 + }, + { + "epoch": 87.99213973799127, + "grad_norm": 0.013733026571571827, + "learning_rate": 0.0006, + "loss": 4.29277229309082, + "step": 6335 + }, + { + "epoch": 88.0, + "grad_norm": 0.014752327464520931, + "learning_rate": 0.0006, + "loss": 4.163867950439453, + "step": 6336 + }, + { + "epoch": 88.0, + "eval_loss": 4.535566329956055, + "eval_runtime": 56.4677, + "eval_samples_per_second": 43.246, + "eval_steps_per_second": 1.364, + "step": 6336 + }, + { + "epoch": 88.01397379912663, + "grad_norm": 0.013958732597529888, + "learning_rate": 0.0006, + "loss": 4.200681209564209, + "step": 6337 + }, + { + "epoch": 88.02794759825328, + "grad_norm": 0.015283545479178429, + "learning_rate": 0.0006, + "loss": 4.1736016273498535, + "step": 6338 + }, + { + "epoch": 88.04192139737991, + "grad_norm": 0.016192059963941574, + "learning_rate": 0.0006, + "loss": 4.040589809417725, + "step": 6339 + }, + { + "epoch": 88.05589519650655, + "grad_norm": 0.018096154555678368, + "learning_rate": 0.0006, + "loss": 4.146696090698242, + "step": 6340 + }, + { + "epoch": 88.06986899563319, + "grad_norm": 0.01797664538025856, + "learning_rate": 0.0006, + "loss": 4.007248878479004, + "step": 6341 + }, + { + "epoch": 88.08384279475983, + "grad_norm": 0.016275612637400627, + "learning_rate": 0.0006, + "loss": 4.193206787109375, + "step": 6342 + }, + { + "epoch": 88.09781659388646, + "grad_norm": 0.014210258610546589, + "learning_rate": 0.0006, + "loss": 4.064548492431641, + "step": 6343 + }, + { + "epoch": 88.1117903930131, + "grad_norm": 0.016085151582956314, + "learning_rate": 0.0006, + "loss": 4.003668785095215, + "step": 6344 + }, + { + "epoch": 88.12576419213974, + "grad_norm": 0.01930483989417553, + "learning_rate": 0.0006, + "loss": 4.236578941345215, + "step": 6345 + }, + { + "epoch": 88.13973799126637, + "grad_norm": 0.017140906304121017, + "learning_rate": 0.0006, + "loss": 3.9593005180358887, + "step": 6346 + }, + { + "epoch": 88.15371179039302, + "grad_norm": 0.01521560549736023, + "learning_rate": 0.0006, + "loss": 4.137644290924072, + "step": 6347 + }, + { + "epoch": 88.16768558951965, + "grad_norm": 0.015788141638040543, + "learning_rate": 0.0006, + "loss": 4.124845027923584, + "step": 6348 + }, + { + "epoch": 88.18165938864628, + "grad_norm": 0.01784597896039486, + "learning_rate": 0.0006, + "loss": 4.139750957489014, + "step": 6349 + }, + { + "epoch": 88.19563318777293, + "grad_norm": 0.016266558319330215, + "learning_rate": 0.0006, + "loss": 4.074851989746094, + "step": 6350 + }, + { + "epoch": 88.20960698689956, + "grad_norm": 0.014024483971297741, + "learning_rate": 0.0006, + "loss": 4.164689064025879, + "step": 6351 + }, + { + "epoch": 88.2235807860262, + "grad_norm": 0.015694420784711838, + "learning_rate": 0.0006, + "loss": 4.219979286193848, + "step": 6352 + }, + { + "epoch": 88.23755458515284, + "grad_norm": 0.014897859655320644, + "learning_rate": 0.0006, + "loss": 4.050738334655762, + "step": 6353 + }, + { + "epoch": 88.25152838427948, + "grad_norm": 0.014827374368906021, + "learning_rate": 0.0006, + "loss": 4.137135028839111, + "step": 6354 + }, + { + "epoch": 88.26550218340611, + "grad_norm": 0.01454879529774189, + "learning_rate": 0.0006, + "loss": 3.9980897903442383, + "step": 6355 + }, + { + "epoch": 88.27947598253274, + "grad_norm": 0.015623947605490685, + "learning_rate": 0.0006, + "loss": 4.226551055908203, + "step": 6356 + }, + { + "epoch": 88.29344978165939, + "grad_norm": 0.015527035109698772, + "learning_rate": 0.0006, + "loss": 4.11053466796875, + "step": 6357 + }, + { + "epoch": 88.30742358078602, + "grad_norm": 0.015242574736475945, + "learning_rate": 0.0006, + "loss": 4.105966567993164, + "step": 6358 + }, + { + "epoch": 88.32139737991267, + "grad_norm": 0.01375538855791092, + "learning_rate": 0.0006, + "loss": 4.148453712463379, + "step": 6359 + }, + { + "epoch": 88.3353711790393, + "grad_norm": 0.01595226116478443, + "learning_rate": 0.0006, + "loss": 4.095836162567139, + "step": 6360 + }, + { + "epoch": 88.34934497816595, + "grad_norm": 0.015252740122377872, + "learning_rate": 0.0006, + "loss": 4.136335849761963, + "step": 6361 + }, + { + "epoch": 88.36331877729258, + "grad_norm": 0.014405866153538227, + "learning_rate": 0.0006, + "loss": 4.084691047668457, + "step": 6362 + }, + { + "epoch": 88.37729257641921, + "grad_norm": 0.014755490235984325, + "learning_rate": 0.0006, + "loss": 4.095475673675537, + "step": 6363 + }, + { + "epoch": 88.39126637554585, + "grad_norm": 0.015339708887040615, + "learning_rate": 0.0006, + "loss": 4.163136959075928, + "step": 6364 + }, + { + "epoch": 88.40524017467249, + "grad_norm": 0.014667164534330368, + "learning_rate": 0.0006, + "loss": 4.062265396118164, + "step": 6365 + }, + { + "epoch": 88.41921397379913, + "grad_norm": 0.014588423073291779, + "learning_rate": 0.0006, + "loss": 4.13359260559082, + "step": 6366 + }, + { + "epoch": 88.43318777292576, + "grad_norm": 0.014005377888679504, + "learning_rate": 0.0006, + "loss": 4.219711780548096, + "step": 6367 + }, + { + "epoch": 88.44716157205241, + "grad_norm": 0.012769930064678192, + "learning_rate": 0.0006, + "loss": 4.114658355712891, + "step": 6368 + }, + { + "epoch": 88.46113537117904, + "grad_norm": 0.013632693327963352, + "learning_rate": 0.0006, + "loss": 4.0675225257873535, + "step": 6369 + }, + { + "epoch": 88.47510917030567, + "grad_norm": 0.013829804956912994, + "learning_rate": 0.0006, + "loss": 4.1125969886779785, + "step": 6370 + }, + { + "epoch": 88.48908296943232, + "grad_norm": 0.013882109895348549, + "learning_rate": 0.0006, + "loss": 4.1627092361450195, + "step": 6371 + }, + { + "epoch": 88.50305676855895, + "grad_norm": 0.014023828320205212, + "learning_rate": 0.0006, + "loss": 4.01038932800293, + "step": 6372 + }, + { + "epoch": 88.5170305676856, + "grad_norm": 0.014840706251561642, + "learning_rate": 0.0006, + "loss": 4.117877960205078, + "step": 6373 + }, + { + "epoch": 88.53100436681223, + "grad_norm": 0.014165930449962616, + "learning_rate": 0.0006, + "loss": 4.1864237785339355, + "step": 6374 + }, + { + "epoch": 88.54497816593886, + "grad_norm": 0.014390201307833195, + "learning_rate": 0.0006, + "loss": 4.1044602394104, + "step": 6375 + }, + { + "epoch": 88.5589519650655, + "grad_norm": 0.015037395060062408, + "learning_rate": 0.0006, + "loss": 4.101107120513916, + "step": 6376 + }, + { + "epoch": 88.57292576419214, + "grad_norm": 0.01671452447772026, + "learning_rate": 0.0006, + "loss": 4.170581817626953, + "step": 6377 + }, + { + "epoch": 88.58689956331878, + "grad_norm": 0.01993076130747795, + "learning_rate": 0.0006, + "loss": 4.261412620544434, + "step": 6378 + }, + { + "epoch": 88.60087336244541, + "grad_norm": 0.023721016943454742, + "learning_rate": 0.0006, + "loss": 4.199363708496094, + "step": 6379 + }, + { + "epoch": 88.61484716157206, + "grad_norm": 0.023726163432002068, + "learning_rate": 0.0006, + "loss": 4.061285018920898, + "step": 6380 + }, + { + "epoch": 88.62882096069869, + "grad_norm": 0.018352877348661423, + "learning_rate": 0.0006, + "loss": 4.0721282958984375, + "step": 6381 + }, + { + "epoch": 88.64279475982532, + "grad_norm": 0.01679830253124237, + "learning_rate": 0.0006, + "loss": 4.1717939376831055, + "step": 6382 + }, + { + "epoch": 88.65676855895197, + "grad_norm": 0.017193637788295746, + "learning_rate": 0.0006, + "loss": 4.230783939361572, + "step": 6383 + }, + { + "epoch": 88.6707423580786, + "grad_norm": 0.01581464521586895, + "learning_rate": 0.0006, + "loss": 4.029906272888184, + "step": 6384 + }, + { + "epoch": 88.68471615720524, + "grad_norm": 0.014049704186618328, + "learning_rate": 0.0006, + "loss": 4.192776679992676, + "step": 6385 + }, + { + "epoch": 88.69868995633188, + "grad_norm": 0.017214450985193253, + "learning_rate": 0.0006, + "loss": 4.248916149139404, + "step": 6386 + }, + { + "epoch": 88.71266375545852, + "grad_norm": 0.0173103716224432, + "learning_rate": 0.0006, + "loss": 4.124347686767578, + "step": 6387 + }, + { + "epoch": 88.72663755458515, + "grad_norm": 0.015745092183351517, + "learning_rate": 0.0006, + "loss": 4.192878246307373, + "step": 6388 + }, + { + "epoch": 88.74061135371178, + "grad_norm": 0.013750772923231125, + "learning_rate": 0.0006, + "loss": 4.1564483642578125, + "step": 6389 + }, + { + "epoch": 88.75458515283843, + "grad_norm": 0.014969858340919018, + "learning_rate": 0.0006, + "loss": 4.0980424880981445, + "step": 6390 + }, + { + "epoch": 88.76855895196506, + "grad_norm": 0.01287173479795456, + "learning_rate": 0.0006, + "loss": 4.145712852478027, + "step": 6391 + }, + { + "epoch": 88.78253275109171, + "grad_norm": 0.012716300785541534, + "learning_rate": 0.0006, + "loss": 4.214226722717285, + "step": 6392 + }, + { + "epoch": 88.79650655021834, + "grad_norm": 0.013855204917490482, + "learning_rate": 0.0006, + "loss": 4.1667256355285645, + "step": 6393 + }, + { + "epoch": 88.81048034934499, + "grad_norm": 0.013650864362716675, + "learning_rate": 0.0006, + "loss": 4.161942481994629, + "step": 6394 + }, + { + "epoch": 88.82445414847162, + "grad_norm": 0.013780126348137856, + "learning_rate": 0.0006, + "loss": 4.087825775146484, + "step": 6395 + }, + { + "epoch": 88.83842794759825, + "grad_norm": 0.014029948972165585, + "learning_rate": 0.0006, + "loss": 4.142813205718994, + "step": 6396 + }, + { + "epoch": 88.8524017467249, + "grad_norm": 0.01322692260146141, + "learning_rate": 0.0006, + "loss": 4.106776714324951, + "step": 6397 + }, + { + "epoch": 88.86637554585153, + "grad_norm": 0.013522760011255741, + "learning_rate": 0.0006, + "loss": 4.090996265411377, + "step": 6398 + }, + { + "epoch": 88.88034934497817, + "grad_norm": 0.014474684372544289, + "learning_rate": 0.0006, + "loss": 4.124624252319336, + "step": 6399 + }, + { + "epoch": 88.8943231441048, + "grad_norm": 0.014490422792732716, + "learning_rate": 0.0006, + "loss": 4.065431594848633, + "step": 6400 + }, + { + "epoch": 88.90829694323143, + "grad_norm": 0.015379710122942924, + "learning_rate": 0.0006, + "loss": 4.207939624786377, + "step": 6401 + }, + { + "epoch": 88.92227074235808, + "grad_norm": 0.015914618968963623, + "learning_rate": 0.0006, + "loss": 4.041152000427246, + "step": 6402 + }, + { + "epoch": 88.93624454148471, + "grad_norm": 0.01480048056691885, + "learning_rate": 0.0006, + "loss": 4.138754844665527, + "step": 6403 + }, + { + "epoch": 88.95021834061136, + "grad_norm": 0.013734851032495499, + "learning_rate": 0.0006, + "loss": 4.25323486328125, + "step": 6404 + }, + { + "epoch": 88.96419213973799, + "grad_norm": 0.01452311035245657, + "learning_rate": 0.0006, + "loss": 4.115228176116943, + "step": 6405 + }, + { + "epoch": 88.97816593886463, + "grad_norm": 0.016197681427001953, + "learning_rate": 0.0006, + "loss": 4.099531173706055, + "step": 6406 + }, + { + "epoch": 88.99213973799127, + "grad_norm": 0.017169523984193802, + "learning_rate": 0.0006, + "loss": 4.182981491088867, + "step": 6407 + }, + { + "epoch": 89.0, + "grad_norm": 0.017855646088719368, + "learning_rate": 0.0006, + "loss": 4.1470136642456055, + "step": 6408 + }, + { + "epoch": 89.0, + "eval_loss": 4.6035614013671875, + "eval_runtime": 57.0625, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6408 + }, + { + "epoch": 89.01397379912663, + "grad_norm": 0.014834101311862469, + "learning_rate": 0.0006, + "loss": 4.159243583679199, + "step": 6409 + }, + { + "epoch": 89.02794759825328, + "grad_norm": 0.015316772274672985, + "learning_rate": 0.0006, + "loss": 4.0709123611450195, + "step": 6410 + }, + { + "epoch": 89.04192139737991, + "grad_norm": 0.01576591096818447, + "learning_rate": 0.0006, + "loss": 4.0921196937561035, + "step": 6411 + }, + { + "epoch": 89.05589519650655, + "grad_norm": 0.015948045998811722, + "learning_rate": 0.0006, + "loss": 4.181877136230469, + "step": 6412 + }, + { + "epoch": 89.06986899563319, + "grad_norm": 0.016612380743026733, + "learning_rate": 0.0006, + "loss": 4.081249237060547, + "step": 6413 + }, + { + "epoch": 89.08384279475983, + "grad_norm": 0.017455685883760452, + "learning_rate": 0.0006, + "loss": 4.175990104675293, + "step": 6414 + }, + { + "epoch": 89.09781659388646, + "grad_norm": 0.015040521509945393, + "learning_rate": 0.0006, + "loss": 4.121378421783447, + "step": 6415 + }, + { + "epoch": 89.1117903930131, + "grad_norm": 0.014513209462165833, + "learning_rate": 0.0006, + "loss": 3.996107816696167, + "step": 6416 + }, + { + "epoch": 89.12576419213974, + "grad_norm": 0.015455638989806175, + "learning_rate": 0.0006, + "loss": 4.073675155639648, + "step": 6417 + }, + { + "epoch": 89.13973799126637, + "grad_norm": 0.015468129888176918, + "learning_rate": 0.0006, + "loss": 4.130242824554443, + "step": 6418 + }, + { + "epoch": 89.15371179039302, + "grad_norm": 0.015284297987818718, + "learning_rate": 0.0006, + "loss": 4.18555212020874, + "step": 6419 + }, + { + "epoch": 89.16768558951965, + "grad_norm": 0.01476874016225338, + "learning_rate": 0.0006, + "loss": 4.134175777435303, + "step": 6420 + }, + { + "epoch": 89.18165938864628, + "grad_norm": 0.015836201608181, + "learning_rate": 0.0006, + "loss": 4.094529151916504, + "step": 6421 + }, + { + "epoch": 89.19563318777293, + "grad_norm": 0.0161928478628397, + "learning_rate": 0.0006, + "loss": 4.12647819519043, + "step": 6422 + }, + { + "epoch": 89.20960698689956, + "grad_norm": 0.015680814161896706, + "learning_rate": 0.0006, + "loss": 3.9620392322540283, + "step": 6423 + }, + { + "epoch": 89.2235807860262, + "grad_norm": 0.01584681309759617, + "learning_rate": 0.0006, + "loss": 4.049304008483887, + "step": 6424 + }, + { + "epoch": 89.23755458515284, + "grad_norm": 0.01452907919883728, + "learning_rate": 0.0006, + "loss": 4.064027786254883, + "step": 6425 + }, + { + "epoch": 89.25152838427948, + "grad_norm": 0.014002552255988121, + "learning_rate": 0.0006, + "loss": 4.151862621307373, + "step": 6426 + }, + { + "epoch": 89.26550218340611, + "grad_norm": 0.014658625237643719, + "learning_rate": 0.0006, + "loss": 4.1672210693359375, + "step": 6427 + }, + { + "epoch": 89.27947598253274, + "grad_norm": 0.015021766535937786, + "learning_rate": 0.0006, + "loss": 3.9742586612701416, + "step": 6428 + }, + { + "epoch": 89.29344978165939, + "grad_norm": 0.014681321568787098, + "learning_rate": 0.0006, + "loss": 4.024159908294678, + "step": 6429 + }, + { + "epoch": 89.30742358078602, + "grad_norm": 0.014161549508571625, + "learning_rate": 0.0006, + "loss": 4.105640888214111, + "step": 6430 + }, + { + "epoch": 89.32139737991267, + "grad_norm": 0.014801453799009323, + "learning_rate": 0.0006, + "loss": 4.123049736022949, + "step": 6431 + }, + { + "epoch": 89.3353711790393, + "grad_norm": 0.014307181350886822, + "learning_rate": 0.0006, + "loss": 4.086954593658447, + "step": 6432 + }, + { + "epoch": 89.34934497816595, + "grad_norm": 0.013299369253218174, + "learning_rate": 0.0006, + "loss": 4.109533786773682, + "step": 6433 + }, + { + "epoch": 89.36331877729258, + "grad_norm": 0.014388669282197952, + "learning_rate": 0.0006, + "loss": 4.222370147705078, + "step": 6434 + }, + { + "epoch": 89.37729257641921, + "grad_norm": 0.014501879923045635, + "learning_rate": 0.0006, + "loss": 4.224178314208984, + "step": 6435 + }, + { + "epoch": 89.39126637554585, + "grad_norm": 0.014109021984040737, + "learning_rate": 0.0006, + "loss": 4.112702369689941, + "step": 6436 + }, + { + "epoch": 89.40524017467249, + "grad_norm": 0.013521002605557442, + "learning_rate": 0.0006, + "loss": 4.010124683380127, + "step": 6437 + }, + { + "epoch": 89.41921397379913, + "grad_norm": 0.013288144953548908, + "learning_rate": 0.0006, + "loss": 4.127592086791992, + "step": 6438 + }, + { + "epoch": 89.43318777292576, + "grad_norm": 0.013882486149668694, + "learning_rate": 0.0006, + "loss": 4.031871318817139, + "step": 6439 + }, + { + "epoch": 89.44716157205241, + "grad_norm": 0.01289335172623396, + "learning_rate": 0.0006, + "loss": 4.0757551193237305, + "step": 6440 + }, + { + "epoch": 89.46113537117904, + "grad_norm": 0.01356302760541439, + "learning_rate": 0.0006, + "loss": 4.075678825378418, + "step": 6441 + }, + { + "epoch": 89.47510917030567, + "grad_norm": 0.01447434164583683, + "learning_rate": 0.0006, + "loss": 4.093453407287598, + "step": 6442 + }, + { + "epoch": 89.48908296943232, + "grad_norm": 0.014443481341004372, + "learning_rate": 0.0006, + "loss": 4.075372219085693, + "step": 6443 + }, + { + "epoch": 89.50305676855895, + "grad_norm": 0.01314218994230032, + "learning_rate": 0.0006, + "loss": 4.171533584594727, + "step": 6444 + }, + { + "epoch": 89.5170305676856, + "grad_norm": 0.01348346658051014, + "learning_rate": 0.0006, + "loss": 4.227236270904541, + "step": 6445 + }, + { + "epoch": 89.53100436681223, + "grad_norm": 0.013370679691433907, + "learning_rate": 0.0006, + "loss": 3.973062515258789, + "step": 6446 + }, + { + "epoch": 89.54497816593886, + "grad_norm": 0.01410642359405756, + "learning_rate": 0.0006, + "loss": 4.231175422668457, + "step": 6447 + }, + { + "epoch": 89.5589519650655, + "grad_norm": 0.01419337373226881, + "learning_rate": 0.0006, + "loss": 4.145868301391602, + "step": 6448 + }, + { + "epoch": 89.57292576419214, + "grad_norm": 0.013691053725779057, + "learning_rate": 0.0006, + "loss": 4.175131797790527, + "step": 6449 + }, + { + "epoch": 89.58689956331878, + "grad_norm": 0.013714492321014404, + "learning_rate": 0.0006, + "loss": 4.251893997192383, + "step": 6450 + }, + { + "epoch": 89.60087336244541, + "grad_norm": 0.01393197476863861, + "learning_rate": 0.0006, + "loss": 4.047079086303711, + "step": 6451 + }, + { + "epoch": 89.61484716157206, + "grad_norm": 0.013957309536635876, + "learning_rate": 0.0006, + "loss": 4.128530502319336, + "step": 6452 + }, + { + "epoch": 89.62882096069869, + "grad_norm": 0.013346714898943901, + "learning_rate": 0.0006, + "loss": 4.2296929359436035, + "step": 6453 + }, + { + "epoch": 89.64279475982532, + "grad_norm": 0.014243829064071178, + "learning_rate": 0.0006, + "loss": 4.128233909606934, + "step": 6454 + }, + { + "epoch": 89.65676855895197, + "grad_norm": 0.01606106385588646, + "learning_rate": 0.0006, + "loss": 4.202202320098877, + "step": 6455 + }, + { + "epoch": 89.6707423580786, + "grad_norm": 0.016074348241090775, + "learning_rate": 0.0006, + "loss": 4.215323448181152, + "step": 6456 + }, + { + "epoch": 89.68471615720524, + "grad_norm": 0.01611054316163063, + "learning_rate": 0.0006, + "loss": 4.096758842468262, + "step": 6457 + }, + { + "epoch": 89.69868995633188, + "grad_norm": 0.013788217678666115, + "learning_rate": 0.0006, + "loss": 4.098560810089111, + "step": 6458 + }, + { + "epoch": 89.71266375545852, + "grad_norm": 0.014408969320356846, + "learning_rate": 0.0006, + "loss": 4.107398509979248, + "step": 6459 + }, + { + "epoch": 89.72663755458515, + "grad_norm": 0.014765267260372639, + "learning_rate": 0.0006, + "loss": 4.057718276977539, + "step": 6460 + }, + { + "epoch": 89.74061135371178, + "grad_norm": 0.017395280301570892, + "learning_rate": 0.0006, + "loss": 4.100795269012451, + "step": 6461 + }, + { + "epoch": 89.75458515283843, + "grad_norm": 0.016821149736642838, + "learning_rate": 0.0006, + "loss": 4.153731346130371, + "step": 6462 + }, + { + "epoch": 89.76855895196506, + "grad_norm": 0.015633124858140945, + "learning_rate": 0.0006, + "loss": 4.061585903167725, + "step": 6463 + }, + { + "epoch": 89.78253275109171, + "grad_norm": 0.01451497245579958, + "learning_rate": 0.0006, + "loss": 4.1624884605407715, + "step": 6464 + }, + { + "epoch": 89.79650655021834, + "grad_norm": 0.013868392445147038, + "learning_rate": 0.0006, + "loss": 4.178536415100098, + "step": 6465 + }, + { + "epoch": 89.81048034934499, + "grad_norm": 0.014259875752031803, + "learning_rate": 0.0006, + "loss": 4.073114395141602, + "step": 6466 + }, + { + "epoch": 89.82445414847162, + "grad_norm": 0.013651594519615173, + "learning_rate": 0.0006, + "loss": 4.050265312194824, + "step": 6467 + }, + { + "epoch": 89.83842794759825, + "grad_norm": 0.014063458889722824, + "learning_rate": 0.0006, + "loss": 4.2014689445495605, + "step": 6468 + }, + { + "epoch": 89.8524017467249, + "grad_norm": 0.014179578050971031, + "learning_rate": 0.0006, + "loss": 4.110550880432129, + "step": 6469 + }, + { + "epoch": 89.86637554585153, + "grad_norm": 0.013616854324936867, + "learning_rate": 0.0006, + "loss": 4.1776580810546875, + "step": 6470 + }, + { + "epoch": 89.88034934497817, + "grad_norm": 0.014351330697536469, + "learning_rate": 0.0006, + "loss": 4.058526992797852, + "step": 6471 + }, + { + "epoch": 89.8943231441048, + "grad_norm": 0.0141244325786829, + "learning_rate": 0.0006, + "loss": 4.068673610687256, + "step": 6472 + }, + { + "epoch": 89.90829694323143, + "grad_norm": 0.01552131213247776, + "learning_rate": 0.0006, + "loss": 4.086787223815918, + "step": 6473 + }, + { + "epoch": 89.92227074235808, + "grad_norm": 0.014614894054830074, + "learning_rate": 0.0006, + "loss": 4.2190656661987305, + "step": 6474 + }, + { + "epoch": 89.93624454148471, + "grad_norm": 0.014369758777320385, + "learning_rate": 0.0006, + "loss": 4.117021560668945, + "step": 6475 + }, + { + "epoch": 89.95021834061136, + "grad_norm": 0.015352046117186546, + "learning_rate": 0.0006, + "loss": 4.167664527893066, + "step": 6476 + }, + { + "epoch": 89.96419213973799, + "grad_norm": 0.014314834959805012, + "learning_rate": 0.0006, + "loss": 4.160182952880859, + "step": 6477 + }, + { + "epoch": 89.97816593886463, + "grad_norm": 0.013511341996490955, + "learning_rate": 0.0006, + "loss": 4.096034049987793, + "step": 6478 + }, + { + "epoch": 89.99213973799127, + "grad_norm": 0.01308482326567173, + "learning_rate": 0.0006, + "loss": 4.062689781188965, + "step": 6479 + }, + { + "epoch": 90.0, + "grad_norm": 0.01442806702107191, + "learning_rate": 0.0006, + "loss": 4.1955885887146, + "step": 6480 + }, + { + "epoch": 90.0, + "eval_loss": 4.644347667694092, + "eval_runtime": 56.8677, + "eval_samples_per_second": 42.942, + "eval_steps_per_second": 1.354, + "step": 6480 + }, + { + "epoch": 90.01397379912663, + "grad_norm": 0.014839007519185543, + "learning_rate": 0.0006, + "loss": 4.171868324279785, + "step": 6481 + }, + { + "epoch": 90.02794759825328, + "grad_norm": 0.01476900465786457, + "learning_rate": 0.0006, + "loss": 4.119754791259766, + "step": 6482 + }, + { + "epoch": 90.04192139737991, + "grad_norm": 0.015179301612079144, + "learning_rate": 0.0006, + "loss": 4.06740140914917, + "step": 6483 + }, + { + "epoch": 90.05589519650655, + "grad_norm": 0.015552980825304985, + "learning_rate": 0.0006, + "loss": 4.112268447875977, + "step": 6484 + }, + { + "epoch": 90.06986899563319, + "grad_norm": 0.014930234290659428, + "learning_rate": 0.0006, + "loss": 4.071866989135742, + "step": 6485 + }, + { + "epoch": 90.08384279475983, + "grad_norm": 0.014124148525297642, + "learning_rate": 0.0006, + "loss": 4.013731956481934, + "step": 6486 + }, + { + "epoch": 90.09781659388646, + "grad_norm": 0.014204252511262894, + "learning_rate": 0.0006, + "loss": 4.133620738983154, + "step": 6487 + }, + { + "epoch": 90.1117903930131, + "grad_norm": 0.014845543541014194, + "learning_rate": 0.0006, + "loss": 4.122956275939941, + "step": 6488 + }, + { + "epoch": 90.12576419213974, + "grad_norm": 0.0146744754165411, + "learning_rate": 0.0006, + "loss": 4.087121963500977, + "step": 6489 + }, + { + "epoch": 90.13973799126637, + "grad_norm": 0.016007190570235252, + "learning_rate": 0.0006, + "loss": 3.98740816116333, + "step": 6490 + }, + { + "epoch": 90.15371179039302, + "grad_norm": 0.018072661012411118, + "learning_rate": 0.0006, + "loss": 4.1155009269714355, + "step": 6491 + }, + { + "epoch": 90.16768558951965, + "grad_norm": 0.016868166625499725, + "learning_rate": 0.0006, + "loss": 4.123727321624756, + "step": 6492 + }, + { + "epoch": 90.18165938864628, + "grad_norm": 0.017091726884245872, + "learning_rate": 0.0006, + "loss": 4.1405229568481445, + "step": 6493 + }, + { + "epoch": 90.19563318777293, + "grad_norm": 0.018329855054616928, + "learning_rate": 0.0006, + "loss": 4.093900680541992, + "step": 6494 + }, + { + "epoch": 90.20960698689956, + "grad_norm": 0.018906457349658012, + "learning_rate": 0.0006, + "loss": 4.0552849769592285, + "step": 6495 + }, + { + "epoch": 90.2235807860262, + "grad_norm": 0.017275869846343994, + "learning_rate": 0.0006, + "loss": 4.1310625076293945, + "step": 6496 + }, + { + "epoch": 90.23755458515284, + "grad_norm": 0.015029897913336754, + "learning_rate": 0.0006, + "loss": 4.100166320800781, + "step": 6497 + }, + { + "epoch": 90.25152838427948, + "grad_norm": 0.014891199767589569, + "learning_rate": 0.0006, + "loss": 4.1314802169799805, + "step": 6498 + }, + { + "epoch": 90.26550218340611, + "grad_norm": 0.014877690002322197, + "learning_rate": 0.0006, + "loss": 4.076098442077637, + "step": 6499 + }, + { + "epoch": 90.27947598253274, + "grad_norm": 0.014670864678919315, + "learning_rate": 0.0006, + "loss": 4.066137313842773, + "step": 6500 + }, + { + "epoch": 90.29344978165939, + "grad_norm": 0.01444234699010849, + "learning_rate": 0.0006, + "loss": 4.013591766357422, + "step": 6501 + }, + { + "epoch": 90.30742358078602, + "grad_norm": 0.01339263841509819, + "learning_rate": 0.0006, + "loss": 4.117959499359131, + "step": 6502 + }, + { + "epoch": 90.32139737991267, + "grad_norm": 0.014753810130059719, + "learning_rate": 0.0006, + "loss": 4.042947769165039, + "step": 6503 + }, + { + "epoch": 90.3353711790393, + "grad_norm": 0.014947964809834957, + "learning_rate": 0.0006, + "loss": 4.112810134887695, + "step": 6504 + }, + { + "epoch": 90.34934497816595, + "grad_norm": 0.015155353583395481, + "learning_rate": 0.0006, + "loss": 4.213214874267578, + "step": 6505 + }, + { + "epoch": 90.36331877729258, + "grad_norm": 0.015502216294407845, + "learning_rate": 0.0006, + "loss": 4.127619743347168, + "step": 6506 + }, + { + "epoch": 90.37729257641921, + "grad_norm": 0.015377781353890896, + "learning_rate": 0.0006, + "loss": 4.135862350463867, + "step": 6507 + }, + { + "epoch": 90.39126637554585, + "grad_norm": 0.01662006787955761, + "learning_rate": 0.0006, + "loss": 4.076877593994141, + "step": 6508 + }, + { + "epoch": 90.40524017467249, + "grad_norm": 0.016324395313858986, + "learning_rate": 0.0006, + "loss": 4.214895725250244, + "step": 6509 + }, + { + "epoch": 90.41921397379913, + "grad_norm": 0.016099024564027786, + "learning_rate": 0.0006, + "loss": 4.139717102050781, + "step": 6510 + }, + { + "epoch": 90.43318777292576, + "grad_norm": 0.014403503388166428, + "learning_rate": 0.0006, + "loss": 4.040343284606934, + "step": 6511 + }, + { + "epoch": 90.44716157205241, + "grad_norm": 0.01413416862487793, + "learning_rate": 0.0006, + "loss": 4.087702751159668, + "step": 6512 + }, + { + "epoch": 90.46113537117904, + "grad_norm": 0.01603989489376545, + "learning_rate": 0.0006, + "loss": 4.173558712005615, + "step": 6513 + }, + { + "epoch": 90.47510917030567, + "grad_norm": 0.016846206039190292, + "learning_rate": 0.0006, + "loss": 4.1955461502075195, + "step": 6514 + }, + { + "epoch": 90.48908296943232, + "grad_norm": 0.01652812585234642, + "learning_rate": 0.0006, + "loss": 4.061838150024414, + "step": 6515 + }, + { + "epoch": 90.50305676855895, + "grad_norm": 0.015100528486073017, + "learning_rate": 0.0006, + "loss": 4.041808605194092, + "step": 6516 + }, + { + "epoch": 90.5170305676856, + "grad_norm": 0.013867822475731373, + "learning_rate": 0.0006, + "loss": 4.12662410736084, + "step": 6517 + }, + { + "epoch": 90.53100436681223, + "grad_norm": 0.013684107922017574, + "learning_rate": 0.0006, + "loss": 4.1254048347473145, + "step": 6518 + }, + { + "epoch": 90.54497816593886, + "grad_norm": 0.014200013130903244, + "learning_rate": 0.0006, + "loss": 4.183216094970703, + "step": 6519 + }, + { + "epoch": 90.5589519650655, + "grad_norm": 0.013707022182643414, + "learning_rate": 0.0006, + "loss": 4.07176399230957, + "step": 6520 + }, + { + "epoch": 90.57292576419214, + "grad_norm": 0.013996582478284836, + "learning_rate": 0.0006, + "loss": 4.006575584411621, + "step": 6521 + }, + { + "epoch": 90.58689956331878, + "grad_norm": 0.014886210672557354, + "learning_rate": 0.0006, + "loss": 4.118956565856934, + "step": 6522 + }, + { + "epoch": 90.60087336244541, + "grad_norm": 0.014022685587406158, + "learning_rate": 0.0006, + "loss": 3.96986722946167, + "step": 6523 + }, + { + "epoch": 90.61484716157206, + "grad_norm": 0.014547634869813919, + "learning_rate": 0.0006, + "loss": 4.122886657714844, + "step": 6524 + }, + { + "epoch": 90.62882096069869, + "grad_norm": 0.013682851567864418, + "learning_rate": 0.0006, + "loss": 3.988292694091797, + "step": 6525 + }, + { + "epoch": 90.64279475982532, + "grad_norm": 0.014016421511769295, + "learning_rate": 0.0006, + "loss": 4.110270977020264, + "step": 6526 + }, + { + "epoch": 90.65676855895197, + "grad_norm": 0.013392233289778233, + "learning_rate": 0.0006, + "loss": 4.065059661865234, + "step": 6527 + }, + { + "epoch": 90.6707423580786, + "grad_norm": 0.012107795104384422, + "learning_rate": 0.0006, + "loss": 4.034260272979736, + "step": 6528 + }, + { + "epoch": 90.68471615720524, + "grad_norm": 0.013198604807257652, + "learning_rate": 0.0006, + "loss": 4.162083625793457, + "step": 6529 + }, + { + "epoch": 90.69868995633188, + "grad_norm": 0.012380579486489296, + "learning_rate": 0.0006, + "loss": 4.040456295013428, + "step": 6530 + }, + { + "epoch": 90.71266375545852, + "grad_norm": 0.012784135527908802, + "learning_rate": 0.0006, + "loss": 4.0823798179626465, + "step": 6531 + }, + { + "epoch": 90.72663755458515, + "grad_norm": 0.012920220382511616, + "learning_rate": 0.0006, + "loss": 4.181064128875732, + "step": 6532 + }, + { + "epoch": 90.74061135371178, + "grad_norm": 0.01338814664632082, + "learning_rate": 0.0006, + "loss": 4.111660003662109, + "step": 6533 + }, + { + "epoch": 90.75458515283843, + "grad_norm": 0.013628040440380573, + "learning_rate": 0.0006, + "loss": 4.237852573394775, + "step": 6534 + }, + { + "epoch": 90.76855895196506, + "grad_norm": 0.013384854421019554, + "learning_rate": 0.0006, + "loss": 4.121862888336182, + "step": 6535 + }, + { + "epoch": 90.78253275109171, + "grad_norm": 0.014726929366588593, + "learning_rate": 0.0006, + "loss": 4.1695990562438965, + "step": 6536 + }, + { + "epoch": 90.79650655021834, + "grad_norm": 0.0161006897687912, + "learning_rate": 0.0006, + "loss": 4.17794132232666, + "step": 6537 + }, + { + "epoch": 90.81048034934499, + "grad_norm": 0.016761884093284607, + "learning_rate": 0.0006, + "loss": 4.093764781951904, + "step": 6538 + }, + { + "epoch": 90.82445414847162, + "grad_norm": 0.017246825620532036, + "learning_rate": 0.0006, + "loss": 3.9382331371307373, + "step": 6539 + }, + { + "epoch": 90.83842794759825, + "grad_norm": 0.018141290172934532, + "learning_rate": 0.0006, + "loss": 3.973609209060669, + "step": 6540 + }, + { + "epoch": 90.8524017467249, + "grad_norm": 0.018129250034689903, + "learning_rate": 0.0006, + "loss": 4.227414131164551, + "step": 6541 + }, + { + "epoch": 90.86637554585153, + "grad_norm": 0.018234573304653168, + "learning_rate": 0.0006, + "loss": 4.151870250701904, + "step": 6542 + }, + { + "epoch": 90.88034934497817, + "grad_norm": 0.01560172438621521, + "learning_rate": 0.0006, + "loss": 4.127397537231445, + "step": 6543 + }, + { + "epoch": 90.8943231441048, + "grad_norm": 0.014367681927978992, + "learning_rate": 0.0006, + "loss": 4.17820930480957, + "step": 6544 + }, + { + "epoch": 90.90829694323143, + "grad_norm": 0.016071319580078125, + "learning_rate": 0.0006, + "loss": 4.059854507446289, + "step": 6545 + }, + { + "epoch": 90.92227074235808, + "grad_norm": 0.01589169353246689, + "learning_rate": 0.0006, + "loss": 4.282748222351074, + "step": 6546 + }, + { + "epoch": 90.93624454148471, + "grad_norm": 0.013639772310853004, + "learning_rate": 0.0006, + "loss": 4.147368907928467, + "step": 6547 + }, + { + "epoch": 90.95021834061136, + "grad_norm": 0.014248427003622055, + "learning_rate": 0.0006, + "loss": 4.240248680114746, + "step": 6548 + }, + { + "epoch": 90.96419213973799, + "grad_norm": 0.014600320719182491, + "learning_rate": 0.0006, + "loss": 4.122123718261719, + "step": 6549 + }, + { + "epoch": 90.97816593886463, + "grad_norm": 0.014371675439178944, + "learning_rate": 0.0006, + "loss": 4.013302803039551, + "step": 6550 + }, + { + "epoch": 90.99213973799127, + "grad_norm": 0.013133526779711246, + "learning_rate": 0.0006, + "loss": 4.097136497497559, + "step": 6551 + }, + { + "epoch": 91.0, + "grad_norm": 0.015599353238940239, + "learning_rate": 0.0006, + "loss": 4.198554039001465, + "step": 6552 + }, + { + "epoch": 91.0, + "eval_loss": 4.603907585144043, + "eval_runtime": 56.9418, + "eval_samples_per_second": 42.886, + "eval_steps_per_second": 1.352, + "step": 6552 + }, + { + "epoch": 91.01397379912663, + "grad_norm": 0.015917908400297165, + "learning_rate": 0.0006, + "loss": 4.133877754211426, + "step": 6553 + }, + { + "epoch": 91.02794759825328, + "grad_norm": 0.015485800802707672, + "learning_rate": 0.0006, + "loss": 4.111617088317871, + "step": 6554 + }, + { + "epoch": 91.04192139737991, + "grad_norm": 0.012969603762030602, + "learning_rate": 0.0006, + "loss": 4.064828395843506, + "step": 6555 + }, + { + "epoch": 91.05589519650655, + "grad_norm": 0.014482097700238228, + "learning_rate": 0.0006, + "loss": 4.101710319519043, + "step": 6556 + }, + { + "epoch": 91.06986899563319, + "grad_norm": 0.017242740839719772, + "learning_rate": 0.0006, + "loss": 4.015342712402344, + "step": 6557 + }, + { + "epoch": 91.08384279475983, + "grad_norm": 0.017100440338253975, + "learning_rate": 0.0006, + "loss": 3.9801602363586426, + "step": 6558 + }, + { + "epoch": 91.09781659388646, + "grad_norm": 0.015969131141901016, + "learning_rate": 0.0006, + "loss": 4.049958229064941, + "step": 6559 + }, + { + "epoch": 91.1117903930131, + "grad_norm": 0.016018927097320557, + "learning_rate": 0.0006, + "loss": 4.094139099121094, + "step": 6560 + }, + { + "epoch": 91.12576419213974, + "grad_norm": 0.018131926655769348, + "learning_rate": 0.0006, + "loss": 3.9499974250793457, + "step": 6561 + }, + { + "epoch": 91.13973799126637, + "grad_norm": 0.019425220787525177, + "learning_rate": 0.0006, + "loss": 3.9910552501678467, + "step": 6562 + }, + { + "epoch": 91.15371179039302, + "grad_norm": 0.01908663846552372, + "learning_rate": 0.0006, + "loss": 4.032179832458496, + "step": 6563 + }, + { + "epoch": 91.16768558951965, + "grad_norm": 0.01871207356452942, + "learning_rate": 0.0006, + "loss": 4.112380027770996, + "step": 6564 + }, + { + "epoch": 91.18165938864628, + "grad_norm": 0.020867953076958656, + "learning_rate": 0.0006, + "loss": 4.06105899810791, + "step": 6565 + }, + { + "epoch": 91.19563318777293, + "grad_norm": 0.019053412601351738, + "learning_rate": 0.0006, + "loss": 3.992504596710205, + "step": 6566 + }, + { + "epoch": 91.20960698689956, + "grad_norm": 0.016084402799606323, + "learning_rate": 0.0006, + "loss": 4.100559234619141, + "step": 6567 + }, + { + "epoch": 91.2235807860262, + "grad_norm": 0.016857590526342392, + "learning_rate": 0.0006, + "loss": 4.1567277908325195, + "step": 6568 + }, + { + "epoch": 91.23755458515284, + "grad_norm": 0.01702957972884178, + "learning_rate": 0.0006, + "loss": 4.234771728515625, + "step": 6569 + }, + { + "epoch": 91.25152838427948, + "grad_norm": 0.016869494691491127, + "learning_rate": 0.0006, + "loss": 3.9931387901306152, + "step": 6570 + }, + { + "epoch": 91.26550218340611, + "grad_norm": 0.01696581020951271, + "learning_rate": 0.0006, + "loss": 4.184908866882324, + "step": 6571 + }, + { + "epoch": 91.27947598253274, + "grad_norm": 0.014982126653194427, + "learning_rate": 0.0006, + "loss": 4.036668300628662, + "step": 6572 + }, + { + "epoch": 91.29344978165939, + "grad_norm": 0.014363301917910576, + "learning_rate": 0.0006, + "loss": 4.074010848999023, + "step": 6573 + }, + { + "epoch": 91.30742358078602, + "grad_norm": 0.015492337755858898, + "learning_rate": 0.0006, + "loss": 4.127626419067383, + "step": 6574 + }, + { + "epoch": 91.32139737991267, + "grad_norm": 0.015375050716102123, + "learning_rate": 0.0006, + "loss": 4.159229755401611, + "step": 6575 + }, + { + "epoch": 91.3353711790393, + "grad_norm": 0.015068565495312214, + "learning_rate": 0.0006, + "loss": 4.172623634338379, + "step": 6576 + }, + { + "epoch": 91.34934497816595, + "grad_norm": 0.0140256742015481, + "learning_rate": 0.0006, + "loss": 4.207065105438232, + "step": 6577 + }, + { + "epoch": 91.36331877729258, + "grad_norm": 0.014220152981579304, + "learning_rate": 0.0006, + "loss": 4.131444931030273, + "step": 6578 + }, + { + "epoch": 91.37729257641921, + "grad_norm": 0.0143149234354496, + "learning_rate": 0.0006, + "loss": 4.065218448638916, + "step": 6579 + }, + { + "epoch": 91.39126637554585, + "grad_norm": 0.013149659149348736, + "learning_rate": 0.0006, + "loss": 4.106743335723877, + "step": 6580 + }, + { + "epoch": 91.40524017467249, + "grad_norm": 0.014550330117344856, + "learning_rate": 0.0006, + "loss": 4.123722553253174, + "step": 6581 + }, + { + "epoch": 91.41921397379913, + "grad_norm": 0.014611710794270039, + "learning_rate": 0.0006, + "loss": 4.079610824584961, + "step": 6582 + }, + { + "epoch": 91.43318777292576, + "grad_norm": 0.013369769789278507, + "learning_rate": 0.0006, + "loss": 4.133187770843506, + "step": 6583 + }, + { + "epoch": 91.44716157205241, + "grad_norm": 0.013243545778095722, + "learning_rate": 0.0006, + "loss": 4.042607307434082, + "step": 6584 + }, + { + "epoch": 91.46113537117904, + "grad_norm": 0.013388266786932945, + "learning_rate": 0.0006, + "loss": 4.14863395690918, + "step": 6585 + }, + { + "epoch": 91.47510917030567, + "grad_norm": 0.014333273284137249, + "learning_rate": 0.0006, + "loss": 4.109699726104736, + "step": 6586 + }, + { + "epoch": 91.48908296943232, + "grad_norm": 0.012742757797241211, + "learning_rate": 0.0006, + "loss": 4.126582145690918, + "step": 6587 + }, + { + "epoch": 91.50305676855895, + "grad_norm": 0.012880304828286171, + "learning_rate": 0.0006, + "loss": 4.205300331115723, + "step": 6588 + }, + { + "epoch": 91.5170305676856, + "grad_norm": 0.013896043412387371, + "learning_rate": 0.0006, + "loss": 4.132510185241699, + "step": 6589 + }, + { + "epoch": 91.53100436681223, + "grad_norm": 0.014361281879246235, + "learning_rate": 0.0006, + "loss": 4.1487932205200195, + "step": 6590 + }, + { + "epoch": 91.54497816593886, + "grad_norm": 0.015137741342186928, + "learning_rate": 0.0006, + "loss": 4.1564531326293945, + "step": 6591 + }, + { + "epoch": 91.5589519650655, + "grad_norm": 0.014660682529211044, + "learning_rate": 0.0006, + "loss": 4.056064128875732, + "step": 6592 + }, + { + "epoch": 91.57292576419214, + "grad_norm": 0.014888476580381393, + "learning_rate": 0.0006, + "loss": 4.0633344650268555, + "step": 6593 + }, + { + "epoch": 91.58689956331878, + "grad_norm": 0.012162433005869389, + "learning_rate": 0.0006, + "loss": 4.026293754577637, + "step": 6594 + }, + { + "epoch": 91.60087336244541, + "grad_norm": 0.012892271392047405, + "learning_rate": 0.0006, + "loss": 4.267086029052734, + "step": 6595 + }, + { + "epoch": 91.61484716157206, + "grad_norm": 0.01450759544968605, + "learning_rate": 0.0006, + "loss": 4.149359703063965, + "step": 6596 + }, + { + "epoch": 91.62882096069869, + "grad_norm": 0.01419843453913927, + "learning_rate": 0.0006, + "loss": 4.094183921813965, + "step": 6597 + }, + { + "epoch": 91.64279475982532, + "grad_norm": 0.014401616528630257, + "learning_rate": 0.0006, + "loss": 4.125172138214111, + "step": 6598 + }, + { + "epoch": 91.65676855895197, + "grad_norm": 0.013945513404905796, + "learning_rate": 0.0006, + "loss": 4.146623134613037, + "step": 6599 + }, + { + "epoch": 91.6707423580786, + "grad_norm": 0.013516047969460487, + "learning_rate": 0.0006, + "loss": 4.160526275634766, + "step": 6600 + }, + { + "epoch": 91.68471615720524, + "grad_norm": 0.013262378051877022, + "learning_rate": 0.0006, + "loss": 4.074923038482666, + "step": 6601 + }, + { + "epoch": 91.69868995633188, + "grad_norm": 0.014365669339895248, + "learning_rate": 0.0006, + "loss": 4.118094444274902, + "step": 6602 + }, + { + "epoch": 91.71266375545852, + "grad_norm": 0.016713734716176987, + "learning_rate": 0.0006, + "loss": 4.021799087524414, + "step": 6603 + }, + { + "epoch": 91.72663755458515, + "grad_norm": 0.01988680474460125, + "learning_rate": 0.0006, + "loss": 4.238436222076416, + "step": 6604 + }, + { + "epoch": 91.74061135371178, + "grad_norm": 0.022749733179807663, + "learning_rate": 0.0006, + "loss": 4.010470390319824, + "step": 6605 + }, + { + "epoch": 91.75458515283843, + "grad_norm": 0.022533461451530457, + "learning_rate": 0.0006, + "loss": 4.152201175689697, + "step": 6606 + }, + { + "epoch": 91.76855895196506, + "grad_norm": 0.02021082676947117, + "learning_rate": 0.0006, + "loss": 4.120572090148926, + "step": 6607 + }, + { + "epoch": 91.78253275109171, + "grad_norm": 0.016298197209835052, + "learning_rate": 0.0006, + "loss": 4.115815162658691, + "step": 6608 + }, + { + "epoch": 91.79650655021834, + "grad_norm": 0.017562786117196083, + "learning_rate": 0.0006, + "loss": 4.109724044799805, + "step": 6609 + }, + { + "epoch": 91.81048034934499, + "grad_norm": 0.018153153359889984, + "learning_rate": 0.0006, + "loss": 4.143759727478027, + "step": 6610 + }, + { + "epoch": 91.82445414847162, + "grad_norm": 0.014762306585907936, + "learning_rate": 0.0006, + "loss": 4.25596809387207, + "step": 6611 + }, + { + "epoch": 91.83842794759825, + "grad_norm": 0.015074343420565128, + "learning_rate": 0.0006, + "loss": 4.207723140716553, + "step": 6612 + }, + { + "epoch": 91.8524017467249, + "grad_norm": 0.015028579160571098, + "learning_rate": 0.0006, + "loss": 4.01950216293335, + "step": 6613 + }, + { + "epoch": 91.86637554585153, + "grad_norm": 0.013996795751154423, + "learning_rate": 0.0006, + "loss": 4.132997989654541, + "step": 6614 + }, + { + "epoch": 91.88034934497817, + "grad_norm": 0.014978236518800259, + "learning_rate": 0.0006, + "loss": 4.132402420043945, + "step": 6615 + }, + { + "epoch": 91.8943231441048, + "grad_norm": 0.013111517764627934, + "learning_rate": 0.0006, + "loss": 4.1223649978637695, + "step": 6616 + }, + { + "epoch": 91.90829694323143, + "grad_norm": 0.013872234150767326, + "learning_rate": 0.0006, + "loss": 4.096590042114258, + "step": 6617 + }, + { + "epoch": 91.92227074235808, + "grad_norm": 0.013642470352351665, + "learning_rate": 0.0006, + "loss": 4.236227035522461, + "step": 6618 + }, + { + "epoch": 91.93624454148471, + "grad_norm": 0.014684873633086681, + "learning_rate": 0.0006, + "loss": 4.037961959838867, + "step": 6619 + }, + { + "epoch": 91.95021834061136, + "grad_norm": 0.014498206786811352, + "learning_rate": 0.0006, + "loss": 4.043325424194336, + "step": 6620 + }, + { + "epoch": 91.96419213973799, + "grad_norm": 0.012768656946718693, + "learning_rate": 0.0006, + "loss": 4.138803482055664, + "step": 6621 + }, + { + "epoch": 91.97816593886463, + "grad_norm": 0.012155905365943909, + "learning_rate": 0.0006, + "loss": 4.163424968719482, + "step": 6622 + }, + { + "epoch": 91.99213973799127, + "grad_norm": 0.012584930285811424, + "learning_rate": 0.0006, + "loss": 4.086629867553711, + "step": 6623 + }, + { + "epoch": 92.0, + "grad_norm": 0.01578151248395443, + "learning_rate": 0.0006, + "loss": 4.108386993408203, + "step": 6624 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.824044292425646e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6624/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6624/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6624/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-6912/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-6912/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e4c07da00f264e39f747e36434e38eec1c1c2cb --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f71ef97c4ea096ec954b9050ef9e62a495733b2317bbf71f968b5a2413758c9 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6912/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d15bceb7ef69347bf8cf9394a5d2bd6e5910741f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:999438d24e47e87ad0493588d6f4cf83f9e0860905cd03e026fcb40d7c1a8a1e +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..c3ef1d6a71abfca102fd404f57354f3ca09ee701 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac169ba05fadeecfe00246f94826dca416d3fc2b1287c215fafb508d79aec1a2 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..5fbf0d7eae7fc8d4dda5a5846601c548e141e1d0 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40d6b8065c0a4634082fe5db4e072a73cbf9df22d0f465f7470ee490e85b8f06 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..0c1428cff0ef70846cc7092b335c53bf4fc384b0 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b339702460b81003d8a5ee7890a1776405165efac6a55c01b2b259ef1c27513 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..cd8a97b2a3ecba19c6ed2d04fbb06d29237b4ec3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef39912fd9c5e4d2a49b252fb2e37c1d5bb81162d1ad99afecfcd996b765fec9 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-6912/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..80771355cefe444a789dc3e83678b2d55d3b8db3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d0b661f1b18caa6118d27fef199882abc2f72f0aaf9b16c81f9627cbcf08e00 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-6912/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..ca31af02fdcf45c3b7dcdf6824df426dec965568 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/trainer_state.json @@ -0,0 +1,49178 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 96.0, + "eval_steps": 500, + "global_step": 6912, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + }, + { + "epoch": 76.0, + "eval_loss": 4.689640998840332, + "eval_runtime": 56.9095, + "eval_samples_per_second": 42.91, + "eval_steps_per_second": 1.353, + "step": 5472 + }, + { + "epoch": 76.01397379912663, + "grad_norm": 0.015351055189967155, + "learning_rate": 0.0006, + "loss": 4.348268985748291, + "step": 5473 + }, + { + "epoch": 76.02794759825328, + "grad_norm": 0.017164217308163643, + "learning_rate": 0.0006, + "loss": 4.244204998016357, + "step": 5474 + }, + { + "epoch": 76.04192139737991, + "grad_norm": 0.021624932065606117, + "learning_rate": 0.0006, + "loss": 4.311168670654297, + "step": 5475 + }, + { + "epoch": 76.05589519650655, + "grad_norm": 0.028712468221783638, + "learning_rate": 0.0006, + "loss": 4.322287082672119, + "step": 5476 + }, + { + "epoch": 76.06986899563319, + "grad_norm": 0.027915487065911293, + "learning_rate": 0.0006, + "loss": 4.381207466125488, + "step": 5477 + }, + { + "epoch": 76.08384279475983, + "grad_norm": 0.017955303192138672, + "learning_rate": 0.0006, + "loss": 4.2635016441345215, + "step": 5478 + }, + { + "epoch": 76.09781659388646, + "grad_norm": 0.019113514572381973, + "learning_rate": 0.0006, + "loss": 4.399471282958984, + "step": 5479 + }, + { + "epoch": 76.1117903930131, + "grad_norm": 0.017502542585134506, + "learning_rate": 0.0006, + "loss": 4.321204662322998, + "step": 5480 + }, + { + "epoch": 76.12576419213974, + "grad_norm": 0.01852525770664215, + "learning_rate": 0.0006, + "loss": 4.217884063720703, + "step": 5481 + }, + { + "epoch": 76.13973799126637, + "grad_norm": 0.019836828112602234, + "learning_rate": 0.0006, + "loss": 4.278532028198242, + "step": 5482 + }, + { + "epoch": 76.15371179039302, + "grad_norm": 0.02101975679397583, + "learning_rate": 0.0006, + "loss": 4.246720314025879, + "step": 5483 + }, + { + "epoch": 76.16768558951965, + "grad_norm": 0.021202990785241127, + "learning_rate": 0.0006, + "loss": 4.517517566680908, + "step": 5484 + }, + { + "epoch": 76.18165938864628, + "grad_norm": 0.021428769454360008, + "learning_rate": 0.0006, + "loss": 4.3373517990112305, + "step": 5485 + }, + { + "epoch": 76.19563318777293, + "grad_norm": 0.018263190984725952, + "learning_rate": 0.0006, + "loss": 4.402856826782227, + "step": 5486 + }, + { + "epoch": 76.20960698689956, + "grad_norm": 0.019025536254048347, + "learning_rate": 0.0006, + "loss": 4.314227104187012, + "step": 5487 + }, + { + "epoch": 76.2235807860262, + "grad_norm": 0.020918630063533783, + "learning_rate": 0.0006, + "loss": 4.454369068145752, + "step": 5488 + }, + { + "epoch": 76.23755458515284, + "grad_norm": 0.019307617098093033, + "learning_rate": 0.0006, + "loss": 4.247509002685547, + "step": 5489 + }, + { + "epoch": 76.25152838427948, + "grad_norm": 0.014956331811845303, + "learning_rate": 0.0006, + "loss": 4.365175247192383, + "step": 5490 + }, + { + "epoch": 76.26550218340611, + "grad_norm": 0.017213059589266777, + "learning_rate": 0.0006, + "loss": 4.384005069732666, + "step": 5491 + }, + { + "epoch": 76.27947598253274, + "grad_norm": 0.017575597390532494, + "learning_rate": 0.0006, + "loss": 4.41934871673584, + "step": 5492 + }, + { + "epoch": 76.29344978165939, + "grad_norm": 0.015879256650805473, + "learning_rate": 0.0006, + "loss": 4.26675271987915, + "step": 5493 + }, + { + "epoch": 76.30742358078602, + "grad_norm": 0.01564522460103035, + "learning_rate": 0.0006, + "loss": 4.277339935302734, + "step": 5494 + }, + { + "epoch": 76.32139737991267, + "grad_norm": 0.016414586454629898, + "learning_rate": 0.0006, + "loss": 4.4138593673706055, + "step": 5495 + }, + { + "epoch": 76.3353711790393, + "grad_norm": 0.014732337556779385, + "learning_rate": 0.0006, + "loss": 4.281006336212158, + "step": 5496 + }, + { + "epoch": 76.34934497816595, + "grad_norm": 0.015461455099284649, + "learning_rate": 0.0006, + "loss": 4.263422966003418, + "step": 5497 + }, + { + "epoch": 76.36331877729258, + "grad_norm": 0.014529814012348652, + "learning_rate": 0.0006, + "loss": 4.36165189743042, + "step": 5498 + }, + { + "epoch": 76.37729257641921, + "grad_norm": 0.01469483319669962, + "learning_rate": 0.0006, + "loss": 4.309642791748047, + "step": 5499 + }, + { + "epoch": 76.39126637554585, + "grad_norm": 0.016490574926137924, + "learning_rate": 0.0006, + "loss": 4.288150787353516, + "step": 5500 + }, + { + "epoch": 76.40524017467249, + "grad_norm": 0.014593565836548805, + "learning_rate": 0.0006, + "loss": 4.271614074707031, + "step": 5501 + }, + { + "epoch": 76.41921397379913, + "grad_norm": 0.012796535156667233, + "learning_rate": 0.0006, + "loss": 4.420042037963867, + "step": 5502 + }, + { + "epoch": 76.43318777292576, + "grad_norm": 0.016086025163531303, + "learning_rate": 0.0006, + "loss": 4.289943695068359, + "step": 5503 + }, + { + "epoch": 76.44716157205241, + "grad_norm": 0.015439470298588276, + "learning_rate": 0.0006, + "loss": 4.3023271560668945, + "step": 5504 + }, + { + "epoch": 76.46113537117904, + "grad_norm": 0.015066367574036121, + "learning_rate": 0.0006, + "loss": 4.197081565856934, + "step": 5505 + }, + { + "epoch": 76.47510917030567, + "grad_norm": 0.016653263941407204, + "learning_rate": 0.0006, + "loss": 4.333785057067871, + "step": 5506 + }, + { + "epoch": 76.48908296943232, + "grad_norm": 0.014791973866522312, + "learning_rate": 0.0006, + "loss": 4.323811054229736, + "step": 5507 + }, + { + "epoch": 76.50305676855895, + "grad_norm": 0.014694827608764172, + "learning_rate": 0.0006, + "loss": 4.195852279663086, + "step": 5508 + }, + { + "epoch": 76.5170305676856, + "grad_norm": 0.015120396390557289, + "learning_rate": 0.0006, + "loss": 4.42379903793335, + "step": 5509 + }, + { + "epoch": 76.53100436681223, + "grad_norm": 0.014584025368094444, + "learning_rate": 0.0006, + "loss": 4.2702531814575195, + "step": 5510 + }, + { + "epoch": 76.54497816593886, + "grad_norm": 0.013981613330543041, + "learning_rate": 0.0006, + "loss": 4.383029937744141, + "step": 5511 + }, + { + "epoch": 76.5589519650655, + "grad_norm": 0.012930642813444138, + "learning_rate": 0.0006, + "loss": 4.296136856079102, + "step": 5512 + }, + { + "epoch": 76.57292576419214, + "grad_norm": 0.013060989789664745, + "learning_rate": 0.0006, + "loss": 4.346380233764648, + "step": 5513 + }, + { + "epoch": 76.58689956331878, + "grad_norm": 0.013209798373281956, + "learning_rate": 0.0006, + "loss": 4.316067218780518, + "step": 5514 + }, + { + "epoch": 76.60087336244541, + "grad_norm": 0.012423216365277767, + "learning_rate": 0.0006, + "loss": 4.370274543762207, + "step": 5515 + }, + { + "epoch": 76.61484716157206, + "grad_norm": 0.013205411843955517, + "learning_rate": 0.0006, + "loss": 4.429969310760498, + "step": 5516 + }, + { + "epoch": 76.62882096069869, + "grad_norm": 0.013049827888607979, + "learning_rate": 0.0006, + "loss": 4.378549098968506, + "step": 5517 + }, + { + "epoch": 76.64279475982532, + "grad_norm": 0.012075880542397499, + "learning_rate": 0.0006, + "loss": 4.348293304443359, + "step": 5518 + }, + { + "epoch": 76.65676855895197, + "grad_norm": 0.014932847581803799, + "learning_rate": 0.0006, + "loss": 4.282538890838623, + "step": 5519 + }, + { + "epoch": 76.6707423580786, + "grad_norm": 0.015615998767316341, + "learning_rate": 0.0006, + "loss": 4.274641036987305, + "step": 5520 + }, + { + "epoch": 76.68471615720524, + "grad_norm": 0.01424513477832079, + "learning_rate": 0.0006, + "loss": 4.243166923522949, + "step": 5521 + }, + { + "epoch": 76.69868995633188, + "grad_norm": 0.012775546871125698, + "learning_rate": 0.0006, + "loss": 4.335611820220947, + "step": 5522 + }, + { + "epoch": 76.71266375545852, + "grad_norm": 0.014461303129792213, + "learning_rate": 0.0006, + "loss": 4.246976375579834, + "step": 5523 + }, + { + "epoch": 76.72663755458515, + "grad_norm": 0.0145436255261302, + "learning_rate": 0.0006, + "loss": 4.275895118713379, + "step": 5524 + }, + { + "epoch": 76.74061135371178, + "grad_norm": 0.014055085368454456, + "learning_rate": 0.0006, + "loss": 4.331108093261719, + "step": 5525 + }, + { + "epoch": 76.75458515283843, + "grad_norm": 0.011933894827961922, + "learning_rate": 0.0006, + "loss": 4.293721675872803, + "step": 5526 + }, + { + "epoch": 76.76855895196506, + "grad_norm": 0.01277392078191042, + "learning_rate": 0.0006, + "loss": 4.312374114990234, + "step": 5527 + }, + { + "epoch": 76.78253275109171, + "grad_norm": 0.014227936044335365, + "learning_rate": 0.0006, + "loss": 4.238729000091553, + "step": 5528 + }, + { + "epoch": 76.79650655021834, + "grad_norm": 0.01488024927675724, + "learning_rate": 0.0006, + "loss": 4.24587345123291, + "step": 5529 + }, + { + "epoch": 76.81048034934499, + "grad_norm": 0.015341250225901604, + "learning_rate": 0.0006, + "loss": 4.257378578186035, + "step": 5530 + }, + { + "epoch": 76.82445414847162, + "grad_norm": 0.015179437585175037, + "learning_rate": 0.0006, + "loss": 4.2812910079956055, + "step": 5531 + }, + { + "epoch": 76.83842794759825, + "grad_norm": 0.012814724817872047, + "learning_rate": 0.0006, + "loss": 4.2309465408325195, + "step": 5532 + }, + { + "epoch": 76.8524017467249, + "grad_norm": 0.013810716569423676, + "learning_rate": 0.0006, + "loss": 4.325292110443115, + "step": 5533 + }, + { + "epoch": 76.86637554585153, + "grad_norm": 0.017468232661485672, + "learning_rate": 0.0006, + "loss": 4.243788719177246, + "step": 5534 + }, + { + "epoch": 76.88034934497817, + "grad_norm": 0.01739857904613018, + "learning_rate": 0.0006, + "loss": 4.345390796661377, + "step": 5535 + }, + { + "epoch": 76.8943231441048, + "grad_norm": 0.015329061076045036, + "learning_rate": 0.0006, + "loss": 4.415142059326172, + "step": 5536 + }, + { + "epoch": 76.90829694323143, + "grad_norm": 0.01598426327109337, + "learning_rate": 0.0006, + "loss": 4.299558639526367, + "step": 5537 + }, + { + "epoch": 76.92227074235808, + "grad_norm": 0.01621454581618309, + "learning_rate": 0.0006, + "loss": 4.284354209899902, + "step": 5538 + }, + { + "epoch": 76.93624454148471, + "grad_norm": 0.015956968069076538, + "learning_rate": 0.0006, + "loss": 4.293837547302246, + "step": 5539 + }, + { + "epoch": 76.95021834061136, + "grad_norm": 0.01776493526995182, + "learning_rate": 0.0006, + "loss": 4.229729175567627, + "step": 5540 + }, + { + "epoch": 76.96419213973799, + "grad_norm": 0.015317477285861969, + "learning_rate": 0.0006, + "loss": 4.3328447341918945, + "step": 5541 + }, + { + "epoch": 76.97816593886463, + "grad_norm": 0.015140251256525517, + "learning_rate": 0.0006, + "loss": 4.310840606689453, + "step": 5542 + }, + { + "epoch": 76.99213973799127, + "grad_norm": 0.015326812863349915, + "learning_rate": 0.0006, + "loss": 4.230832099914551, + "step": 5543 + }, + { + "epoch": 77.0, + "grad_norm": 0.01607763208448887, + "learning_rate": 0.0006, + "loss": 4.2236328125, + "step": 5544 + }, + { + "epoch": 77.0, + "eval_loss": 4.662221431732178, + "eval_runtime": 56.5194, + "eval_samples_per_second": 43.206, + "eval_steps_per_second": 1.362, + "step": 5544 + }, + { + "epoch": 77.01397379912663, + "grad_norm": 0.014839510433375835, + "learning_rate": 0.0006, + "loss": 4.254111289978027, + "step": 5545 + }, + { + "epoch": 77.02794759825328, + "grad_norm": 0.01519741676747799, + "learning_rate": 0.0006, + "loss": 4.1607184410095215, + "step": 5546 + }, + { + "epoch": 77.04192139737991, + "grad_norm": 0.015472603030502796, + "learning_rate": 0.0006, + "loss": 4.323320388793945, + "step": 5547 + }, + { + "epoch": 77.05589519650655, + "grad_norm": 0.017937948927283287, + "learning_rate": 0.0006, + "loss": 4.308490753173828, + "step": 5548 + }, + { + "epoch": 77.06986899563319, + "grad_norm": 0.020890841260552406, + "learning_rate": 0.0006, + "loss": 4.259101867675781, + "step": 5549 + }, + { + "epoch": 77.08384279475983, + "grad_norm": 0.01892857998609543, + "learning_rate": 0.0006, + "loss": 4.386970520019531, + "step": 5550 + }, + { + "epoch": 77.09781659388646, + "grad_norm": 0.016583630815148354, + "learning_rate": 0.0006, + "loss": 4.233604907989502, + "step": 5551 + }, + { + "epoch": 77.1117903930131, + "grad_norm": 0.016063468530774117, + "learning_rate": 0.0006, + "loss": 4.339544296264648, + "step": 5552 + }, + { + "epoch": 77.12576419213974, + "grad_norm": 0.018326779827475548, + "learning_rate": 0.0006, + "loss": 4.367204666137695, + "step": 5553 + }, + { + "epoch": 77.13973799126637, + "grad_norm": 0.021590933203697205, + "learning_rate": 0.0006, + "loss": 4.352685928344727, + "step": 5554 + }, + { + "epoch": 77.15371179039302, + "grad_norm": 0.019643642008304596, + "learning_rate": 0.0006, + "loss": 4.258535385131836, + "step": 5555 + }, + { + "epoch": 77.16768558951965, + "grad_norm": 0.018476149067282677, + "learning_rate": 0.0006, + "loss": 4.232568740844727, + "step": 5556 + }, + { + "epoch": 77.18165938864628, + "grad_norm": 0.01584405079483986, + "learning_rate": 0.0006, + "loss": 4.258371353149414, + "step": 5557 + }, + { + "epoch": 77.19563318777293, + "grad_norm": 0.01575201191008091, + "learning_rate": 0.0006, + "loss": 4.183218002319336, + "step": 5558 + }, + { + "epoch": 77.20960698689956, + "grad_norm": 0.016871824860572815, + "learning_rate": 0.0006, + "loss": 4.194563865661621, + "step": 5559 + }, + { + "epoch": 77.2235807860262, + "grad_norm": 0.02059333398938179, + "learning_rate": 0.0006, + "loss": 4.195281982421875, + "step": 5560 + }, + { + "epoch": 77.23755458515284, + "grad_norm": 0.022640453651547432, + "learning_rate": 0.0006, + "loss": 4.279615879058838, + "step": 5561 + }, + { + "epoch": 77.25152838427948, + "grad_norm": 0.020335959270596504, + "learning_rate": 0.0006, + "loss": 4.25040864944458, + "step": 5562 + }, + { + "epoch": 77.26550218340611, + "grad_norm": 0.020010793581604958, + "learning_rate": 0.0006, + "loss": 4.291116714477539, + "step": 5563 + }, + { + "epoch": 77.27947598253274, + "grad_norm": 0.02109045907855034, + "learning_rate": 0.0006, + "loss": 4.364088535308838, + "step": 5564 + }, + { + "epoch": 77.29344978165939, + "grad_norm": 0.019519446417689323, + "learning_rate": 0.0006, + "loss": 4.329166412353516, + "step": 5565 + }, + { + "epoch": 77.30742358078602, + "grad_norm": 0.01689804159104824, + "learning_rate": 0.0006, + "loss": 4.244176864624023, + "step": 5566 + }, + { + "epoch": 77.32139737991267, + "grad_norm": 0.01893681474030018, + "learning_rate": 0.0006, + "loss": 4.239068984985352, + "step": 5567 + }, + { + "epoch": 77.3353711790393, + "grad_norm": 0.015998193994164467, + "learning_rate": 0.0006, + "loss": 4.173398017883301, + "step": 5568 + }, + { + "epoch": 77.34934497816595, + "grad_norm": 0.013892280869185925, + "learning_rate": 0.0006, + "loss": 4.287654876708984, + "step": 5569 + }, + { + "epoch": 77.36331877729258, + "grad_norm": 0.017060643061995506, + "learning_rate": 0.0006, + "loss": 4.398447036743164, + "step": 5570 + }, + { + "epoch": 77.37729257641921, + "grad_norm": 0.015168849378824234, + "learning_rate": 0.0006, + "loss": 4.15665864944458, + "step": 5571 + }, + { + "epoch": 77.39126637554585, + "grad_norm": 0.014886523596942425, + "learning_rate": 0.0006, + "loss": 4.196673393249512, + "step": 5572 + }, + { + "epoch": 77.40524017467249, + "grad_norm": 0.016941796988248825, + "learning_rate": 0.0006, + "loss": 4.314562797546387, + "step": 5573 + }, + { + "epoch": 77.41921397379913, + "grad_norm": 0.015353812836110592, + "learning_rate": 0.0006, + "loss": 4.409579277038574, + "step": 5574 + }, + { + "epoch": 77.43318777292576, + "grad_norm": 0.015192852355539799, + "learning_rate": 0.0006, + "loss": 4.392045974731445, + "step": 5575 + }, + { + "epoch": 77.44716157205241, + "grad_norm": 0.014232569374144077, + "learning_rate": 0.0006, + "loss": 4.21639347076416, + "step": 5576 + }, + { + "epoch": 77.46113537117904, + "grad_norm": 0.014453226700425148, + "learning_rate": 0.0006, + "loss": 4.2802276611328125, + "step": 5577 + }, + { + "epoch": 77.47510917030567, + "grad_norm": 0.012881236150860786, + "learning_rate": 0.0006, + "loss": 4.375980854034424, + "step": 5578 + }, + { + "epoch": 77.48908296943232, + "grad_norm": 0.01583976298570633, + "learning_rate": 0.0006, + "loss": 4.264520645141602, + "step": 5579 + }, + { + "epoch": 77.50305676855895, + "grad_norm": 0.014171584509313107, + "learning_rate": 0.0006, + "loss": 4.225279808044434, + "step": 5580 + }, + { + "epoch": 77.5170305676856, + "grad_norm": 0.014715159311890602, + "learning_rate": 0.0006, + "loss": 4.369834899902344, + "step": 5581 + }, + { + "epoch": 77.53100436681223, + "grad_norm": 0.016554079949855804, + "learning_rate": 0.0006, + "loss": 4.2976179122924805, + "step": 5582 + }, + { + "epoch": 77.54497816593886, + "grad_norm": 0.015095747075974941, + "learning_rate": 0.0006, + "loss": 4.294407844543457, + "step": 5583 + }, + { + "epoch": 77.5589519650655, + "grad_norm": 0.012726361863315105, + "learning_rate": 0.0006, + "loss": 4.221401214599609, + "step": 5584 + }, + { + "epoch": 77.57292576419214, + "grad_norm": 0.014999309554696083, + "learning_rate": 0.0006, + "loss": 4.243929386138916, + "step": 5585 + }, + { + "epoch": 77.58689956331878, + "grad_norm": 0.016526037827134132, + "learning_rate": 0.0006, + "loss": 4.374139785766602, + "step": 5586 + }, + { + "epoch": 77.60087336244541, + "grad_norm": 0.01648339442908764, + "learning_rate": 0.0006, + "loss": 4.392425060272217, + "step": 5587 + }, + { + "epoch": 77.61484716157206, + "grad_norm": 0.015269213356077671, + "learning_rate": 0.0006, + "loss": 4.33986759185791, + "step": 5588 + }, + { + "epoch": 77.62882096069869, + "grad_norm": 0.01690507121384144, + "learning_rate": 0.0006, + "loss": 4.307476043701172, + "step": 5589 + }, + { + "epoch": 77.64279475982532, + "grad_norm": 0.016438685357570648, + "learning_rate": 0.0006, + "loss": 4.2406206130981445, + "step": 5590 + }, + { + "epoch": 77.65676855895197, + "grad_norm": 0.0157600287348032, + "learning_rate": 0.0006, + "loss": 4.314070224761963, + "step": 5591 + }, + { + "epoch": 77.6707423580786, + "grad_norm": 0.015127947553992271, + "learning_rate": 0.0006, + "loss": 4.22475528717041, + "step": 5592 + }, + { + "epoch": 77.68471615720524, + "grad_norm": 0.016632677987217903, + "learning_rate": 0.0006, + "loss": 4.26954984664917, + "step": 5593 + }, + { + "epoch": 77.69868995633188, + "grad_norm": 0.013150460086762905, + "learning_rate": 0.0006, + "loss": 4.369131088256836, + "step": 5594 + }, + { + "epoch": 77.71266375545852, + "grad_norm": 0.014513005502521992, + "learning_rate": 0.0006, + "loss": 4.3248796463012695, + "step": 5595 + }, + { + "epoch": 77.72663755458515, + "grad_norm": 0.014090991578996181, + "learning_rate": 0.0006, + "loss": 4.268381118774414, + "step": 5596 + }, + { + "epoch": 77.74061135371178, + "grad_norm": 0.01633097417652607, + "learning_rate": 0.0006, + "loss": 4.238439083099365, + "step": 5597 + }, + { + "epoch": 77.75458515283843, + "grad_norm": 0.01710933819413185, + "learning_rate": 0.0006, + "loss": 4.399697780609131, + "step": 5598 + }, + { + "epoch": 77.76855895196506, + "grad_norm": 0.019251767545938492, + "learning_rate": 0.0006, + "loss": 4.290450096130371, + "step": 5599 + }, + { + "epoch": 77.78253275109171, + "grad_norm": 0.01849171705543995, + "learning_rate": 0.0006, + "loss": 4.267734527587891, + "step": 5600 + }, + { + "epoch": 77.79650655021834, + "grad_norm": 0.014173115603625774, + "learning_rate": 0.0006, + "loss": 4.288799285888672, + "step": 5601 + }, + { + "epoch": 77.81048034934499, + "grad_norm": 0.014572718180716038, + "learning_rate": 0.0006, + "loss": 4.3790435791015625, + "step": 5602 + }, + { + "epoch": 77.82445414847162, + "grad_norm": 0.016375336796045303, + "learning_rate": 0.0006, + "loss": 4.231725692749023, + "step": 5603 + }, + { + "epoch": 77.83842794759825, + "grad_norm": 0.017326701432466507, + "learning_rate": 0.0006, + "loss": 4.211495399475098, + "step": 5604 + }, + { + "epoch": 77.8524017467249, + "grad_norm": 0.016399390995502472, + "learning_rate": 0.0006, + "loss": 4.386699199676514, + "step": 5605 + }, + { + "epoch": 77.86637554585153, + "grad_norm": 0.018269989639520645, + "learning_rate": 0.0006, + "loss": 4.254856109619141, + "step": 5606 + }, + { + "epoch": 77.88034934497817, + "grad_norm": 0.018443435430526733, + "learning_rate": 0.0006, + "loss": 4.391961097717285, + "step": 5607 + }, + { + "epoch": 77.8943231441048, + "grad_norm": 0.01784636080265045, + "learning_rate": 0.0006, + "loss": 4.273881912231445, + "step": 5608 + }, + { + "epoch": 77.90829694323143, + "grad_norm": 0.01782994158565998, + "learning_rate": 0.0006, + "loss": 4.26833963394165, + "step": 5609 + }, + { + "epoch": 77.92227074235808, + "grad_norm": 0.016260741278529167, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 5610 + }, + { + "epoch": 77.93624454148471, + "grad_norm": 0.015482903458178043, + "learning_rate": 0.0006, + "loss": 4.187826633453369, + "step": 5611 + }, + { + "epoch": 77.95021834061136, + "grad_norm": 0.015193257480859756, + "learning_rate": 0.0006, + "loss": 4.286060333251953, + "step": 5612 + }, + { + "epoch": 77.96419213973799, + "grad_norm": 0.015110115520656109, + "learning_rate": 0.0006, + "loss": 4.2828192710876465, + "step": 5613 + }, + { + "epoch": 77.97816593886463, + "grad_norm": 0.015973486006259918, + "learning_rate": 0.0006, + "loss": 4.317939758300781, + "step": 5614 + }, + { + "epoch": 77.99213973799127, + "grad_norm": 0.017363594844937325, + "learning_rate": 0.0006, + "loss": 4.287721633911133, + "step": 5615 + }, + { + "epoch": 78.0, + "grad_norm": 0.015277642756700516, + "learning_rate": 0.0006, + "loss": 4.356618881225586, + "step": 5616 + }, + { + "epoch": 78.0, + "eval_loss": 4.642678737640381, + "eval_runtime": 56.7335, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 5616 + }, + { + "epoch": 78.01397379912663, + "grad_norm": 0.016794051975011826, + "learning_rate": 0.0006, + "loss": 4.279930591583252, + "step": 5617 + }, + { + "epoch": 78.02794759825328, + "grad_norm": 0.01632673107087612, + "learning_rate": 0.0006, + "loss": 4.235943794250488, + "step": 5618 + }, + { + "epoch": 78.04192139737991, + "grad_norm": 0.01437876746058464, + "learning_rate": 0.0006, + "loss": 4.2542243003845215, + "step": 5619 + }, + { + "epoch": 78.05589519650655, + "grad_norm": 0.014993119053542614, + "learning_rate": 0.0006, + "loss": 4.262046813964844, + "step": 5620 + }, + { + "epoch": 78.06986899563319, + "grad_norm": 0.015721119940280914, + "learning_rate": 0.0006, + "loss": 4.173436164855957, + "step": 5621 + }, + { + "epoch": 78.08384279475983, + "grad_norm": 0.01334085687994957, + "learning_rate": 0.0006, + "loss": 4.322175979614258, + "step": 5622 + }, + { + "epoch": 78.09781659388646, + "grad_norm": 0.015561388805508614, + "learning_rate": 0.0006, + "loss": 4.242125988006592, + "step": 5623 + }, + { + "epoch": 78.1117903930131, + "grad_norm": 0.014418602921068668, + "learning_rate": 0.0006, + "loss": 4.266186714172363, + "step": 5624 + }, + { + "epoch": 78.12576419213974, + "grad_norm": 0.013739909045398235, + "learning_rate": 0.0006, + "loss": 4.231633186340332, + "step": 5625 + }, + { + "epoch": 78.13973799126637, + "grad_norm": 0.01402067206799984, + "learning_rate": 0.0006, + "loss": 4.188406944274902, + "step": 5626 + }, + { + "epoch": 78.15371179039302, + "grad_norm": 0.013639481738209724, + "learning_rate": 0.0006, + "loss": 4.1586594581604, + "step": 5627 + }, + { + "epoch": 78.16768558951965, + "grad_norm": 0.012858159840106964, + "learning_rate": 0.0006, + "loss": 4.360448360443115, + "step": 5628 + }, + { + "epoch": 78.18165938864628, + "grad_norm": 0.013089739717543125, + "learning_rate": 0.0006, + "loss": 4.344371795654297, + "step": 5629 + }, + { + "epoch": 78.19563318777293, + "grad_norm": 0.01382314320653677, + "learning_rate": 0.0006, + "loss": 4.20378303527832, + "step": 5630 + }, + { + "epoch": 78.20960698689956, + "grad_norm": 0.014206619001924992, + "learning_rate": 0.0006, + "loss": 4.239891529083252, + "step": 5631 + }, + { + "epoch": 78.2235807860262, + "grad_norm": 0.01472168043255806, + "learning_rate": 0.0006, + "loss": 4.282382488250732, + "step": 5632 + }, + { + "epoch": 78.23755458515284, + "grad_norm": 0.014237025752663612, + "learning_rate": 0.0006, + "loss": 4.241545677185059, + "step": 5633 + }, + { + "epoch": 78.25152838427948, + "grad_norm": 0.013231472112238407, + "learning_rate": 0.0006, + "loss": 4.184896469116211, + "step": 5634 + }, + { + "epoch": 78.26550218340611, + "grad_norm": 0.013538471423089504, + "learning_rate": 0.0006, + "loss": 4.325181007385254, + "step": 5635 + }, + { + "epoch": 78.27947598253274, + "grad_norm": 0.013541961088776588, + "learning_rate": 0.0006, + "loss": 4.368722915649414, + "step": 5636 + }, + { + "epoch": 78.29344978165939, + "grad_norm": 0.013446723110973835, + "learning_rate": 0.0006, + "loss": 4.301422595977783, + "step": 5637 + }, + { + "epoch": 78.30742358078602, + "grad_norm": 0.014753809198737144, + "learning_rate": 0.0006, + "loss": 4.259127616882324, + "step": 5638 + }, + { + "epoch": 78.32139737991267, + "grad_norm": 0.013256008736789227, + "learning_rate": 0.0006, + "loss": 4.284032821655273, + "step": 5639 + }, + { + "epoch": 78.3353711790393, + "grad_norm": 0.014272629283368587, + "learning_rate": 0.0006, + "loss": 4.234989166259766, + "step": 5640 + }, + { + "epoch": 78.34934497816595, + "grad_norm": 0.014756826683878899, + "learning_rate": 0.0006, + "loss": 4.143499374389648, + "step": 5641 + }, + { + "epoch": 78.36331877729258, + "grad_norm": 0.016549205407500267, + "learning_rate": 0.0006, + "loss": 4.311034202575684, + "step": 5642 + }, + { + "epoch": 78.37729257641921, + "grad_norm": 0.017735283821821213, + "learning_rate": 0.0006, + "loss": 4.246767520904541, + "step": 5643 + }, + { + "epoch": 78.39126637554585, + "grad_norm": 0.017370613291859627, + "learning_rate": 0.0006, + "loss": 4.2460832595825195, + "step": 5644 + }, + { + "epoch": 78.40524017467249, + "grad_norm": 0.014642846770584583, + "learning_rate": 0.0006, + "loss": 4.303811550140381, + "step": 5645 + }, + { + "epoch": 78.41921397379913, + "grad_norm": 0.013445726595818996, + "learning_rate": 0.0006, + "loss": 4.25921630859375, + "step": 5646 + }, + { + "epoch": 78.43318777292576, + "grad_norm": 0.014254065230488777, + "learning_rate": 0.0006, + "loss": 4.129446506500244, + "step": 5647 + }, + { + "epoch": 78.44716157205241, + "grad_norm": 0.01883019506931305, + "learning_rate": 0.0006, + "loss": 4.191716194152832, + "step": 5648 + }, + { + "epoch": 78.46113537117904, + "grad_norm": 0.0204437468200922, + "learning_rate": 0.0006, + "loss": 4.333085536956787, + "step": 5649 + }, + { + "epoch": 78.47510917030567, + "grad_norm": 0.01652318798005581, + "learning_rate": 0.0006, + "loss": 4.135462760925293, + "step": 5650 + }, + { + "epoch": 78.48908296943232, + "grad_norm": 0.012753551825881004, + "learning_rate": 0.0006, + "loss": 4.195051193237305, + "step": 5651 + }, + { + "epoch": 78.50305676855895, + "grad_norm": 0.01499118935316801, + "learning_rate": 0.0006, + "loss": 4.157092094421387, + "step": 5652 + }, + { + "epoch": 78.5170305676856, + "grad_norm": 0.015383945778012276, + "learning_rate": 0.0006, + "loss": 4.253818035125732, + "step": 5653 + }, + { + "epoch": 78.53100436681223, + "grad_norm": 0.015721332281827927, + "learning_rate": 0.0006, + "loss": 4.228775978088379, + "step": 5654 + }, + { + "epoch": 78.54497816593886, + "grad_norm": 0.014595463871955872, + "learning_rate": 0.0006, + "loss": 4.135197639465332, + "step": 5655 + }, + { + "epoch": 78.5589519650655, + "grad_norm": 0.015059644356369972, + "learning_rate": 0.0006, + "loss": 4.255614280700684, + "step": 5656 + }, + { + "epoch": 78.57292576419214, + "grad_norm": 0.013833344914019108, + "learning_rate": 0.0006, + "loss": 4.301250457763672, + "step": 5657 + }, + { + "epoch": 78.58689956331878, + "grad_norm": 0.013665376231074333, + "learning_rate": 0.0006, + "loss": 4.1734232902526855, + "step": 5658 + }, + { + "epoch": 78.60087336244541, + "grad_norm": 0.01539086177945137, + "learning_rate": 0.0006, + "loss": 4.221459865570068, + "step": 5659 + }, + { + "epoch": 78.61484716157206, + "grad_norm": 0.013313917443156242, + "learning_rate": 0.0006, + "loss": 4.255619049072266, + "step": 5660 + }, + { + "epoch": 78.62882096069869, + "grad_norm": 0.01447741687297821, + "learning_rate": 0.0006, + "loss": 4.298219680786133, + "step": 5661 + }, + { + "epoch": 78.64279475982532, + "grad_norm": 0.012796915136277676, + "learning_rate": 0.0006, + "loss": 4.2542405128479, + "step": 5662 + }, + { + "epoch": 78.65676855895197, + "grad_norm": 0.014058141969144344, + "learning_rate": 0.0006, + "loss": 4.284477233886719, + "step": 5663 + }, + { + "epoch": 78.6707423580786, + "grad_norm": 0.012705069966614246, + "learning_rate": 0.0006, + "loss": 4.2826666831970215, + "step": 5664 + }, + { + "epoch": 78.68471615720524, + "grad_norm": 0.013245890848338604, + "learning_rate": 0.0006, + "loss": 4.371612548828125, + "step": 5665 + }, + { + "epoch": 78.69868995633188, + "grad_norm": 0.014192677102982998, + "learning_rate": 0.0006, + "loss": 4.267344951629639, + "step": 5666 + }, + { + "epoch": 78.71266375545852, + "grad_norm": 0.015346777625381947, + "learning_rate": 0.0006, + "loss": 4.222797393798828, + "step": 5667 + }, + { + "epoch": 78.72663755458515, + "grad_norm": 0.014068621210753918, + "learning_rate": 0.0006, + "loss": 4.172964572906494, + "step": 5668 + }, + { + "epoch": 78.74061135371178, + "grad_norm": 0.014364945702254772, + "learning_rate": 0.0006, + "loss": 4.259319305419922, + "step": 5669 + }, + { + "epoch": 78.75458515283843, + "grad_norm": 0.017337387427687645, + "learning_rate": 0.0006, + "loss": 4.29533576965332, + "step": 5670 + }, + { + "epoch": 78.76855895196506, + "grad_norm": 0.017416715621948242, + "learning_rate": 0.0006, + "loss": 4.340364456176758, + "step": 5671 + }, + { + "epoch": 78.78253275109171, + "grad_norm": 0.015925846993923187, + "learning_rate": 0.0006, + "loss": 4.177454948425293, + "step": 5672 + }, + { + "epoch": 78.79650655021834, + "grad_norm": 0.013130443170666695, + "learning_rate": 0.0006, + "loss": 4.277454853057861, + "step": 5673 + }, + { + "epoch": 78.81048034934499, + "grad_norm": 0.014628687873482704, + "learning_rate": 0.0006, + "loss": 4.273757457733154, + "step": 5674 + }, + { + "epoch": 78.82445414847162, + "grad_norm": 0.01601659320294857, + "learning_rate": 0.0006, + "loss": 4.281774520874023, + "step": 5675 + }, + { + "epoch": 78.83842794759825, + "grad_norm": 0.015443485230207443, + "learning_rate": 0.0006, + "loss": 4.199937343597412, + "step": 5676 + }, + { + "epoch": 78.8524017467249, + "grad_norm": 0.013934246264398098, + "learning_rate": 0.0006, + "loss": 4.331369876861572, + "step": 5677 + }, + { + "epoch": 78.86637554585153, + "grad_norm": 0.01572083681821823, + "learning_rate": 0.0006, + "loss": 4.2828474044799805, + "step": 5678 + }, + { + "epoch": 78.88034934497817, + "grad_norm": 0.013444120064377785, + "learning_rate": 0.0006, + "loss": 4.183859825134277, + "step": 5679 + }, + { + "epoch": 78.8943231441048, + "grad_norm": 0.01591254211962223, + "learning_rate": 0.0006, + "loss": 4.295256614685059, + "step": 5680 + }, + { + "epoch": 78.90829694323143, + "grad_norm": 0.017556684091687202, + "learning_rate": 0.0006, + "loss": 4.132272720336914, + "step": 5681 + }, + { + "epoch": 78.92227074235808, + "grad_norm": 0.018184322863817215, + "learning_rate": 0.0006, + "loss": 4.251728057861328, + "step": 5682 + }, + { + "epoch": 78.93624454148471, + "grad_norm": 0.016207285225391388, + "learning_rate": 0.0006, + "loss": 4.205544471740723, + "step": 5683 + }, + { + "epoch": 78.95021834061136, + "grad_norm": 0.014478376135230064, + "learning_rate": 0.0006, + "loss": 4.182462215423584, + "step": 5684 + }, + { + "epoch": 78.96419213973799, + "grad_norm": 0.014255235902965069, + "learning_rate": 0.0006, + "loss": 4.253366947174072, + "step": 5685 + }, + { + "epoch": 78.97816593886463, + "grad_norm": 0.01524396613240242, + "learning_rate": 0.0006, + "loss": 4.221786022186279, + "step": 5686 + }, + { + "epoch": 78.99213973799127, + "grad_norm": 0.015057693235576153, + "learning_rate": 0.0006, + "loss": 4.208454608917236, + "step": 5687 + }, + { + "epoch": 79.0, + "grad_norm": 0.015478022396564484, + "learning_rate": 0.0006, + "loss": 4.336383819580078, + "step": 5688 + }, + { + "epoch": 79.0, + "eval_loss": 4.641749382019043, + "eval_runtime": 56.8167, + "eval_samples_per_second": 42.98, + "eval_steps_per_second": 1.355, + "step": 5688 + }, + { + "epoch": 79.01397379912663, + "grad_norm": 0.014143316075205803, + "learning_rate": 0.0006, + "loss": 4.260974884033203, + "step": 5689 + }, + { + "epoch": 79.02794759825328, + "grad_norm": 0.013120883144438267, + "learning_rate": 0.0006, + "loss": 4.139886856079102, + "step": 5690 + }, + { + "epoch": 79.04192139737991, + "grad_norm": 0.015019661746919155, + "learning_rate": 0.0006, + "loss": 4.267004013061523, + "step": 5691 + }, + { + "epoch": 79.05589519650655, + "grad_norm": 0.014891284517943859, + "learning_rate": 0.0006, + "loss": 4.205019950866699, + "step": 5692 + }, + { + "epoch": 79.06986899563319, + "grad_norm": 0.015527728945016861, + "learning_rate": 0.0006, + "loss": 4.1513566970825195, + "step": 5693 + }, + { + "epoch": 79.08384279475983, + "grad_norm": 0.017527181655168533, + "learning_rate": 0.0006, + "loss": 4.148828983306885, + "step": 5694 + }, + { + "epoch": 79.09781659388646, + "grad_norm": 0.01632206328213215, + "learning_rate": 0.0006, + "loss": 4.195789813995361, + "step": 5695 + }, + { + "epoch": 79.1117903930131, + "grad_norm": 0.01590503193438053, + "learning_rate": 0.0006, + "loss": 4.199155330657959, + "step": 5696 + }, + { + "epoch": 79.12576419213974, + "grad_norm": 0.016489997506141663, + "learning_rate": 0.0006, + "loss": 4.155154705047607, + "step": 5697 + }, + { + "epoch": 79.13973799126637, + "grad_norm": 0.0161368977278471, + "learning_rate": 0.0006, + "loss": 4.164384841918945, + "step": 5698 + }, + { + "epoch": 79.15371179039302, + "grad_norm": 0.018574809655547142, + "learning_rate": 0.0006, + "loss": 4.1898193359375, + "step": 5699 + }, + { + "epoch": 79.16768558951965, + "grad_norm": 0.018987983465194702, + "learning_rate": 0.0006, + "loss": 4.219719409942627, + "step": 5700 + }, + { + "epoch": 79.18165938864628, + "grad_norm": 0.01668606884777546, + "learning_rate": 0.0006, + "loss": 4.152734756469727, + "step": 5701 + }, + { + "epoch": 79.19563318777293, + "grad_norm": 0.014046299271285534, + "learning_rate": 0.0006, + "loss": 4.152964115142822, + "step": 5702 + }, + { + "epoch": 79.20960698689956, + "grad_norm": 0.01628032699227333, + "learning_rate": 0.0006, + "loss": 4.299479961395264, + "step": 5703 + }, + { + "epoch": 79.2235807860262, + "grad_norm": 0.017998795956373215, + "learning_rate": 0.0006, + "loss": 4.231203079223633, + "step": 5704 + }, + { + "epoch": 79.23755458515284, + "grad_norm": 0.014799647964537144, + "learning_rate": 0.0006, + "loss": 4.3530192375183105, + "step": 5705 + }, + { + "epoch": 79.25152838427948, + "grad_norm": 0.014640678651630878, + "learning_rate": 0.0006, + "loss": 4.287322998046875, + "step": 5706 + }, + { + "epoch": 79.26550218340611, + "grad_norm": 0.0172616895288229, + "learning_rate": 0.0006, + "loss": 4.148009777069092, + "step": 5707 + }, + { + "epoch": 79.27947598253274, + "grad_norm": 0.01811790093779564, + "learning_rate": 0.0006, + "loss": 4.269433975219727, + "step": 5708 + }, + { + "epoch": 79.29344978165939, + "grad_norm": 0.01764000579714775, + "learning_rate": 0.0006, + "loss": 4.292638778686523, + "step": 5709 + }, + { + "epoch": 79.30742358078602, + "grad_norm": 0.018030935898423195, + "learning_rate": 0.0006, + "loss": 4.267889976501465, + "step": 5710 + }, + { + "epoch": 79.32139737991267, + "grad_norm": 0.017860891297459602, + "learning_rate": 0.0006, + "loss": 4.211110591888428, + "step": 5711 + }, + { + "epoch": 79.3353711790393, + "grad_norm": 0.017295073717832565, + "learning_rate": 0.0006, + "loss": 4.224645614624023, + "step": 5712 + }, + { + "epoch": 79.34934497816595, + "grad_norm": 0.015891101211309433, + "learning_rate": 0.0006, + "loss": 4.224797248840332, + "step": 5713 + }, + { + "epoch": 79.36331877729258, + "grad_norm": 0.014680417254567146, + "learning_rate": 0.0006, + "loss": 4.182242393493652, + "step": 5714 + }, + { + "epoch": 79.37729257641921, + "grad_norm": 0.016621140763163567, + "learning_rate": 0.0006, + "loss": 4.186334609985352, + "step": 5715 + }, + { + "epoch": 79.39126637554585, + "grad_norm": 0.016207249835133553, + "learning_rate": 0.0006, + "loss": 4.277074813842773, + "step": 5716 + }, + { + "epoch": 79.40524017467249, + "grad_norm": 0.015375478193163872, + "learning_rate": 0.0006, + "loss": 4.154317855834961, + "step": 5717 + }, + { + "epoch": 79.41921397379913, + "grad_norm": 0.013453228399157524, + "learning_rate": 0.0006, + "loss": 4.265270233154297, + "step": 5718 + }, + { + "epoch": 79.43318777292576, + "grad_norm": 0.014312694780528545, + "learning_rate": 0.0006, + "loss": 4.232220649719238, + "step": 5719 + }, + { + "epoch": 79.44716157205241, + "grad_norm": 0.014759554527699947, + "learning_rate": 0.0006, + "loss": 4.272275924682617, + "step": 5720 + }, + { + "epoch": 79.46113537117904, + "grad_norm": 0.015885505825281143, + "learning_rate": 0.0006, + "loss": 4.2229719161987305, + "step": 5721 + }, + { + "epoch": 79.47510917030567, + "grad_norm": 0.016516581177711487, + "learning_rate": 0.0006, + "loss": 4.200624465942383, + "step": 5722 + }, + { + "epoch": 79.48908296943232, + "grad_norm": 0.018254555761814117, + "learning_rate": 0.0006, + "loss": 4.19648551940918, + "step": 5723 + }, + { + "epoch": 79.50305676855895, + "grad_norm": 0.01467090006917715, + "learning_rate": 0.0006, + "loss": 4.135190486907959, + "step": 5724 + }, + { + "epoch": 79.5170305676856, + "grad_norm": 0.013474797829985619, + "learning_rate": 0.0006, + "loss": 4.311361312866211, + "step": 5725 + }, + { + "epoch": 79.53100436681223, + "grad_norm": 0.01742764748632908, + "learning_rate": 0.0006, + "loss": 4.1845903396606445, + "step": 5726 + }, + { + "epoch": 79.54497816593886, + "grad_norm": 0.018786882981657982, + "learning_rate": 0.0006, + "loss": 4.163942337036133, + "step": 5727 + }, + { + "epoch": 79.5589519650655, + "grad_norm": 0.017831219360232353, + "learning_rate": 0.0006, + "loss": 4.236685276031494, + "step": 5728 + }, + { + "epoch": 79.57292576419214, + "grad_norm": 0.016334986314177513, + "learning_rate": 0.0006, + "loss": 4.255853652954102, + "step": 5729 + }, + { + "epoch": 79.58689956331878, + "grad_norm": 0.0138487434014678, + "learning_rate": 0.0006, + "loss": 4.227874279022217, + "step": 5730 + }, + { + "epoch": 79.60087336244541, + "grad_norm": 0.01531760673969984, + "learning_rate": 0.0006, + "loss": 4.221432685852051, + "step": 5731 + }, + { + "epoch": 79.61484716157206, + "grad_norm": 0.01603361964225769, + "learning_rate": 0.0006, + "loss": 4.278326511383057, + "step": 5732 + }, + { + "epoch": 79.62882096069869, + "grad_norm": 0.013149895705282688, + "learning_rate": 0.0006, + "loss": 4.2474517822265625, + "step": 5733 + }, + { + "epoch": 79.64279475982532, + "grad_norm": 0.013428786769509315, + "learning_rate": 0.0006, + "loss": 4.1701178550720215, + "step": 5734 + }, + { + "epoch": 79.65676855895197, + "grad_norm": 0.014305585995316505, + "learning_rate": 0.0006, + "loss": 4.250778675079346, + "step": 5735 + }, + { + "epoch": 79.6707423580786, + "grad_norm": 0.014544566161930561, + "learning_rate": 0.0006, + "loss": 4.243236541748047, + "step": 5736 + }, + { + "epoch": 79.68471615720524, + "grad_norm": 0.013524414040148258, + "learning_rate": 0.0006, + "loss": 4.255457878112793, + "step": 5737 + }, + { + "epoch": 79.69868995633188, + "grad_norm": 0.014750641770660877, + "learning_rate": 0.0006, + "loss": 4.202816486358643, + "step": 5738 + }, + { + "epoch": 79.71266375545852, + "grad_norm": 0.015090773813426495, + "learning_rate": 0.0006, + "loss": 4.3053178787231445, + "step": 5739 + }, + { + "epoch": 79.72663755458515, + "grad_norm": 0.01606169529259205, + "learning_rate": 0.0006, + "loss": 4.224017143249512, + "step": 5740 + }, + { + "epoch": 79.74061135371178, + "grad_norm": 0.016303520649671555, + "learning_rate": 0.0006, + "loss": 4.190760612487793, + "step": 5741 + }, + { + "epoch": 79.75458515283843, + "grad_norm": 0.01465672068297863, + "learning_rate": 0.0006, + "loss": 4.24506950378418, + "step": 5742 + }, + { + "epoch": 79.76855895196506, + "grad_norm": 0.012616301886737347, + "learning_rate": 0.0006, + "loss": 4.174039840698242, + "step": 5743 + }, + { + "epoch": 79.78253275109171, + "grad_norm": 0.01358284242451191, + "learning_rate": 0.0006, + "loss": 4.236380577087402, + "step": 5744 + }, + { + "epoch": 79.79650655021834, + "grad_norm": 0.01557177770882845, + "learning_rate": 0.0006, + "loss": 4.102962493896484, + "step": 5745 + }, + { + "epoch": 79.81048034934499, + "grad_norm": 0.015584317035973072, + "learning_rate": 0.0006, + "loss": 4.242603302001953, + "step": 5746 + }, + { + "epoch": 79.82445414847162, + "grad_norm": 0.012597334571182728, + "learning_rate": 0.0006, + "loss": 4.202791690826416, + "step": 5747 + }, + { + "epoch": 79.83842794759825, + "grad_norm": 0.013722898438572884, + "learning_rate": 0.0006, + "loss": 4.204328536987305, + "step": 5748 + }, + { + "epoch": 79.8524017467249, + "grad_norm": 0.014669225551187992, + "learning_rate": 0.0006, + "loss": 4.164331912994385, + "step": 5749 + }, + { + "epoch": 79.86637554585153, + "grad_norm": 0.015012740157544613, + "learning_rate": 0.0006, + "loss": 4.260565280914307, + "step": 5750 + }, + { + "epoch": 79.88034934497817, + "grad_norm": 0.016026340425014496, + "learning_rate": 0.0006, + "loss": 4.299723148345947, + "step": 5751 + }, + { + "epoch": 79.8943231441048, + "grad_norm": 0.01626049168407917, + "learning_rate": 0.0006, + "loss": 4.207252502441406, + "step": 5752 + }, + { + "epoch": 79.90829694323143, + "grad_norm": 0.015415354631841183, + "learning_rate": 0.0006, + "loss": 4.151934623718262, + "step": 5753 + }, + { + "epoch": 79.92227074235808, + "grad_norm": 0.014158242382109165, + "learning_rate": 0.0006, + "loss": 4.3461995124816895, + "step": 5754 + }, + { + "epoch": 79.93624454148471, + "grad_norm": 0.01585628278553486, + "learning_rate": 0.0006, + "loss": 4.334761619567871, + "step": 5755 + }, + { + "epoch": 79.95021834061136, + "grad_norm": 0.013710759580135345, + "learning_rate": 0.0006, + "loss": 4.180401802062988, + "step": 5756 + }, + { + "epoch": 79.96419213973799, + "grad_norm": 0.014159204438328743, + "learning_rate": 0.0006, + "loss": 4.322331428527832, + "step": 5757 + }, + { + "epoch": 79.97816593886463, + "grad_norm": 0.015153195708990097, + "learning_rate": 0.0006, + "loss": 4.287478923797607, + "step": 5758 + }, + { + "epoch": 79.99213973799127, + "grad_norm": 0.014100227504968643, + "learning_rate": 0.0006, + "loss": 4.231807708740234, + "step": 5759 + }, + { + "epoch": 80.0, + "grad_norm": 0.015248682349920273, + "learning_rate": 0.0006, + "loss": 4.261106014251709, + "step": 5760 + }, + { + "epoch": 80.0, + "eval_loss": 4.610264301300049, + "eval_runtime": 56.514, + "eval_samples_per_second": 43.211, + "eval_steps_per_second": 1.362, + "step": 5760 + }, + { + "epoch": 80.01397379912663, + "grad_norm": 0.014730271883308887, + "learning_rate": 0.0006, + "loss": 4.108556747436523, + "step": 5761 + }, + { + "epoch": 80.02794759825328, + "grad_norm": 0.014575716108083725, + "learning_rate": 0.0006, + "loss": 4.169658660888672, + "step": 5762 + }, + { + "epoch": 80.04192139737991, + "grad_norm": 0.014706958085298538, + "learning_rate": 0.0006, + "loss": 4.252662658691406, + "step": 5763 + }, + { + "epoch": 80.05589519650655, + "grad_norm": 0.01561224739998579, + "learning_rate": 0.0006, + "loss": 4.193892955780029, + "step": 5764 + }, + { + "epoch": 80.06986899563319, + "grad_norm": 0.014702324755489826, + "learning_rate": 0.0006, + "loss": 4.1889238357543945, + "step": 5765 + }, + { + "epoch": 80.08384279475983, + "grad_norm": 0.015565741807222366, + "learning_rate": 0.0006, + "loss": 4.243549823760986, + "step": 5766 + }, + { + "epoch": 80.09781659388646, + "grad_norm": 0.01647181063890457, + "learning_rate": 0.0006, + "loss": 4.275446891784668, + "step": 5767 + }, + { + "epoch": 80.1117903930131, + "grad_norm": 0.018177716061472893, + "learning_rate": 0.0006, + "loss": 4.12040376663208, + "step": 5768 + }, + { + "epoch": 80.12576419213974, + "grad_norm": 0.01745191030204296, + "learning_rate": 0.0006, + "loss": 4.140371322631836, + "step": 5769 + }, + { + "epoch": 80.13973799126637, + "grad_norm": 0.01568993367254734, + "learning_rate": 0.0006, + "loss": 4.258280277252197, + "step": 5770 + }, + { + "epoch": 80.15371179039302, + "grad_norm": 0.01636597141623497, + "learning_rate": 0.0006, + "loss": 4.179052352905273, + "step": 5771 + }, + { + "epoch": 80.16768558951965, + "grad_norm": 0.016839100047945976, + "learning_rate": 0.0006, + "loss": 4.127636432647705, + "step": 5772 + }, + { + "epoch": 80.18165938864628, + "grad_norm": 0.015471016988158226, + "learning_rate": 0.0006, + "loss": 4.181325912475586, + "step": 5773 + }, + { + "epoch": 80.19563318777293, + "grad_norm": 0.015836089849472046, + "learning_rate": 0.0006, + "loss": 4.142168998718262, + "step": 5774 + }, + { + "epoch": 80.20960698689956, + "grad_norm": 0.014782954007387161, + "learning_rate": 0.0006, + "loss": 4.308799743652344, + "step": 5775 + }, + { + "epoch": 80.2235807860262, + "grad_norm": 0.014040893875062466, + "learning_rate": 0.0006, + "loss": 4.216936111450195, + "step": 5776 + }, + { + "epoch": 80.23755458515284, + "grad_norm": 0.01420280709862709, + "learning_rate": 0.0006, + "loss": 4.235230445861816, + "step": 5777 + }, + { + "epoch": 80.25152838427948, + "grad_norm": 0.013341007754206657, + "learning_rate": 0.0006, + "loss": 4.369109153747559, + "step": 5778 + }, + { + "epoch": 80.26550218340611, + "grad_norm": 0.014208327978849411, + "learning_rate": 0.0006, + "loss": 4.219112396240234, + "step": 5779 + }, + { + "epoch": 80.27947598253274, + "grad_norm": 0.013825052417814732, + "learning_rate": 0.0006, + "loss": 4.1983489990234375, + "step": 5780 + }, + { + "epoch": 80.29344978165939, + "grad_norm": 0.012916515581309795, + "learning_rate": 0.0006, + "loss": 4.204582691192627, + "step": 5781 + }, + { + "epoch": 80.30742358078602, + "grad_norm": 0.013071243651211262, + "learning_rate": 0.0006, + "loss": 4.17643928527832, + "step": 5782 + }, + { + "epoch": 80.32139737991267, + "grad_norm": 0.015507174655795097, + "learning_rate": 0.0006, + "loss": 4.172976016998291, + "step": 5783 + }, + { + "epoch": 80.3353711790393, + "grad_norm": 0.014967243187129498, + "learning_rate": 0.0006, + "loss": 4.141168594360352, + "step": 5784 + }, + { + "epoch": 80.34934497816595, + "grad_norm": 0.015629924833774567, + "learning_rate": 0.0006, + "loss": 4.253526210784912, + "step": 5785 + }, + { + "epoch": 80.36331877729258, + "grad_norm": 0.01607954502105713, + "learning_rate": 0.0006, + "loss": 4.249395370483398, + "step": 5786 + }, + { + "epoch": 80.37729257641921, + "grad_norm": 0.01500706560909748, + "learning_rate": 0.0006, + "loss": 4.279012680053711, + "step": 5787 + }, + { + "epoch": 80.39126637554585, + "grad_norm": 0.014343900606036186, + "learning_rate": 0.0006, + "loss": 4.268509864807129, + "step": 5788 + }, + { + "epoch": 80.40524017467249, + "grad_norm": 0.014682197012007236, + "learning_rate": 0.0006, + "loss": 4.206334114074707, + "step": 5789 + }, + { + "epoch": 80.41921397379913, + "grad_norm": 0.014989163726568222, + "learning_rate": 0.0006, + "loss": 4.143601417541504, + "step": 5790 + }, + { + "epoch": 80.43318777292576, + "grad_norm": 0.015001608058810234, + "learning_rate": 0.0006, + "loss": 4.164111137390137, + "step": 5791 + }, + { + "epoch": 80.44716157205241, + "grad_norm": 0.015222722664475441, + "learning_rate": 0.0006, + "loss": 4.2369914054870605, + "step": 5792 + }, + { + "epoch": 80.46113537117904, + "grad_norm": 0.014126183465123177, + "learning_rate": 0.0006, + "loss": 4.2209930419921875, + "step": 5793 + }, + { + "epoch": 80.47510917030567, + "grad_norm": 0.013065959326922894, + "learning_rate": 0.0006, + "loss": 4.231184959411621, + "step": 5794 + }, + { + "epoch": 80.48908296943232, + "grad_norm": 0.013833962380886078, + "learning_rate": 0.0006, + "loss": 4.226434707641602, + "step": 5795 + }, + { + "epoch": 80.50305676855895, + "grad_norm": 0.01387125626206398, + "learning_rate": 0.0006, + "loss": 4.241328716278076, + "step": 5796 + }, + { + "epoch": 80.5170305676856, + "grad_norm": 0.014145063236355782, + "learning_rate": 0.0006, + "loss": 4.2268805503845215, + "step": 5797 + }, + { + "epoch": 80.53100436681223, + "grad_norm": 0.012990240007638931, + "learning_rate": 0.0006, + "loss": 4.1718339920043945, + "step": 5798 + }, + { + "epoch": 80.54497816593886, + "grad_norm": 0.014185411855578423, + "learning_rate": 0.0006, + "loss": 4.232397556304932, + "step": 5799 + }, + { + "epoch": 80.5589519650655, + "grad_norm": 0.014260523952543736, + "learning_rate": 0.0006, + "loss": 4.282734394073486, + "step": 5800 + }, + { + "epoch": 80.57292576419214, + "grad_norm": 0.012542120181024075, + "learning_rate": 0.0006, + "loss": 4.134551525115967, + "step": 5801 + }, + { + "epoch": 80.58689956331878, + "grad_norm": 0.013078657910227776, + "learning_rate": 0.0006, + "loss": 4.229815483093262, + "step": 5802 + }, + { + "epoch": 80.60087336244541, + "grad_norm": 0.012821056880056858, + "learning_rate": 0.0006, + "loss": 4.208278179168701, + "step": 5803 + }, + { + "epoch": 80.61484716157206, + "grad_norm": 0.012964180670678616, + "learning_rate": 0.0006, + "loss": 4.188477039337158, + "step": 5804 + }, + { + "epoch": 80.62882096069869, + "grad_norm": 0.01631746254861355, + "learning_rate": 0.0006, + "loss": 4.3214497566223145, + "step": 5805 + }, + { + "epoch": 80.64279475982532, + "grad_norm": 0.01726640947163105, + "learning_rate": 0.0006, + "loss": 4.279359817504883, + "step": 5806 + }, + { + "epoch": 80.65676855895197, + "grad_norm": 0.018778258934617043, + "learning_rate": 0.0006, + "loss": 4.294047832489014, + "step": 5807 + }, + { + "epoch": 80.6707423580786, + "grad_norm": 0.01796831749379635, + "learning_rate": 0.0006, + "loss": 4.314162254333496, + "step": 5808 + }, + { + "epoch": 80.68471615720524, + "grad_norm": 0.016495538875460625, + "learning_rate": 0.0006, + "loss": 4.273871421813965, + "step": 5809 + }, + { + "epoch": 80.69868995633188, + "grad_norm": 0.013590306043624878, + "learning_rate": 0.0006, + "loss": 4.14961051940918, + "step": 5810 + }, + { + "epoch": 80.71266375545852, + "grad_norm": 0.014670426957309246, + "learning_rate": 0.0006, + "loss": 4.211853981018066, + "step": 5811 + }, + { + "epoch": 80.72663755458515, + "grad_norm": 0.01641341857612133, + "learning_rate": 0.0006, + "loss": 4.211948871612549, + "step": 5812 + }, + { + "epoch": 80.74061135371178, + "grad_norm": 0.016365470364689827, + "learning_rate": 0.0006, + "loss": 4.200774192810059, + "step": 5813 + }, + { + "epoch": 80.75458515283843, + "grad_norm": 0.01613970287144184, + "learning_rate": 0.0006, + "loss": 4.288322448730469, + "step": 5814 + }, + { + "epoch": 80.76855895196506, + "grad_norm": 0.016742777079343796, + "learning_rate": 0.0006, + "loss": 4.23801326751709, + "step": 5815 + }, + { + "epoch": 80.78253275109171, + "grad_norm": 0.015351069159805775, + "learning_rate": 0.0006, + "loss": 4.263832092285156, + "step": 5816 + }, + { + "epoch": 80.79650655021834, + "grad_norm": 0.013435336761176586, + "learning_rate": 0.0006, + "loss": 4.160857200622559, + "step": 5817 + }, + { + "epoch": 80.81048034934499, + "grad_norm": 0.01474886853247881, + "learning_rate": 0.0006, + "loss": 4.216025352478027, + "step": 5818 + }, + { + "epoch": 80.82445414847162, + "grad_norm": 0.01586935855448246, + "learning_rate": 0.0006, + "loss": 4.256402969360352, + "step": 5819 + }, + { + "epoch": 80.83842794759825, + "grad_norm": 0.014596572145819664, + "learning_rate": 0.0006, + "loss": 4.142584800720215, + "step": 5820 + }, + { + "epoch": 80.8524017467249, + "grad_norm": 0.01501593366265297, + "learning_rate": 0.0006, + "loss": 4.250462055206299, + "step": 5821 + }, + { + "epoch": 80.86637554585153, + "grad_norm": 0.014773511327803135, + "learning_rate": 0.0006, + "loss": 4.253305912017822, + "step": 5822 + }, + { + "epoch": 80.88034934497817, + "grad_norm": 0.013246326707303524, + "learning_rate": 0.0006, + "loss": 4.170881271362305, + "step": 5823 + }, + { + "epoch": 80.8943231441048, + "grad_norm": 0.013013756833970547, + "learning_rate": 0.0006, + "loss": 4.264721870422363, + "step": 5824 + }, + { + "epoch": 80.90829694323143, + "grad_norm": 0.013155903667211533, + "learning_rate": 0.0006, + "loss": 4.274667739868164, + "step": 5825 + }, + { + "epoch": 80.92227074235808, + "grad_norm": 0.01448952779173851, + "learning_rate": 0.0006, + "loss": 4.24777889251709, + "step": 5826 + }, + { + "epoch": 80.93624454148471, + "grad_norm": 0.01415231078863144, + "learning_rate": 0.0006, + "loss": 4.22014856338501, + "step": 5827 + }, + { + "epoch": 80.95021834061136, + "grad_norm": 0.014796273782849312, + "learning_rate": 0.0006, + "loss": 4.22036600112915, + "step": 5828 + }, + { + "epoch": 80.96419213973799, + "grad_norm": 0.013005326502025127, + "learning_rate": 0.0006, + "loss": 4.276123046875, + "step": 5829 + }, + { + "epoch": 80.97816593886463, + "grad_norm": 0.012851177714765072, + "learning_rate": 0.0006, + "loss": 4.248575687408447, + "step": 5830 + }, + { + "epoch": 80.99213973799127, + "grad_norm": 0.013740002177655697, + "learning_rate": 0.0006, + "loss": 4.051140308380127, + "step": 5831 + }, + { + "epoch": 81.0, + "grad_norm": 0.015296104364097118, + "learning_rate": 0.0006, + "loss": 4.251155853271484, + "step": 5832 + }, + { + "epoch": 81.0, + "eval_loss": 4.594127655029297, + "eval_runtime": 56.0959, + "eval_samples_per_second": 43.533, + "eval_steps_per_second": 1.373, + "step": 5832 + }, + { + "epoch": 81.01397379912663, + "grad_norm": 0.015804558992385864, + "learning_rate": 0.0006, + "loss": 4.152894020080566, + "step": 5833 + }, + { + "epoch": 81.02794759825328, + "grad_norm": 0.017000854015350342, + "learning_rate": 0.0006, + "loss": 4.266067981719971, + "step": 5834 + }, + { + "epoch": 81.04192139737991, + "grad_norm": 0.016351008787751198, + "learning_rate": 0.0006, + "loss": 4.192703723907471, + "step": 5835 + }, + { + "epoch": 81.05589519650655, + "grad_norm": 0.014703895896673203, + "learning_rate": 0.0006, + "loss": 4.23554801940918, + "step": 5836 + }, + { + "epoch": 81.06986899563319, + "grad_norm": 0.01619596965610981, + "learning_rate": 0.0006, + "loss": 4.233000755310059, + "step": 5837 + }, + { + "epoch": 81.08384279475983, + "grad_norm": 0.01815493032336235, + "learning_rate": 0.0006, + "loss": 4.200490951538086, + "step": 5838 + }, + { + "epoch": 81.09781659388646, + "grad_norm": 0.018305106088519096, + "learning_rate": 0.0006, + "loss": 4.154242515563965, + "step": 5839 + }, + { + "epoch": 81.1117903930131, + "grad_norm": 0.0200856514275074, + "learning_rate": 0.0006, + "loss": 4.25820779800415, + "step": 5840 + }, + { + "epoch": 81.12576419213974, + "grad_norm": 0.02101440541446209, + "learning_rate": 0.0006, + "loss": 4.168209075927734, + "step": 5841 + }, + { + "epoch": 81.13973799126637, + "grad_norm": 0.022829681634902954, + "learning_rate": 0.0006, + "loss": 4.25560998916626, + "step": 5842 + }, + { + "epoch": 81.15371179039302, + "grad_norm": 0.0195082426071167, + "learning_rate": 0.0006, + "loss": 4.2035627365112305, + "step": 5843 + }, + { + "epoch": 81.16768558951965, + "grad_norm": 0.01716248318552971, + "learning_rate": 0.0006, + "loss": 4.11322021484375, + "step": 5844 + }, + { + "epoch": 81.18165938864628, + "grad_norm": 0.02034664712846279, + "learning_rate": 0.0006, + "loss": 4.179389953613281, + "step": 5845 + }, + { + "epoch": 81.19563318777293, + "grad_norm": 0.023578710854053497, + "learning_rate": 0.0006, + "loss": 4.265695571899414, + "step": 5846 + }, + { + "epoch": 81.20960698689956, + "grad_norm": 0.019244182854890823, + "learning_rate": 0.0006, + "loss": 4.292043209075928, + "step": 5847 + }, + { + "epoch": 81.2235807860262, + "grad_norm": 0.01577291637659073, + "learning_rate": 0.0006, + "loss": 4.250776290893555, + "step": 5848 + }, + { + "epoch": 81.23755458515284, + "grad_norm": 0.01617090217769146, + "learning_rate": 0.0006, + "loss": 4.193098068237305, + "step": 5849 + }, + { + "epoch": 81.25152838427948, + "grad_norm": 0.016090184450149536, + "learning_rate": 0.0006, + "loss": 4.112357139587402, + "step": 5850 + }, + { + "epoch": 81.26550218340611, + "grad_norm": 0.01441365759819746, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 5851 + }, + { + "epoch": 81.27947598253274, + "grad_norm": 0.014122162945568562, + "learning_rate": 0.0006, + "loss": 4.276175022125244, + "step": 5852 + }, + { + "epoch": 81.29344978165939, + "grad_norm": 0.014859487302601337, + "learning_rate": 0.0006, + "loss": 4.094699859619141, + "step": 5853 + }, + { + "epoch": 81.30742358078602, + "grad_norm": 0.013624267652630806, + "learning_rate": 0.0006, + "loss": 4.2648773193359375, + "step": 5854 + }, + { + "epoch": 81.32139737991267, + "grad_norm": 0.01417913381010294, + "learning_rate": 0.0006, + "loss": 4.228567123413086, + "step": 5855 + }, + { + "epoch": 81.3353711790393, + "grad_norm": 0.014560805633664131, + "learning_rate": 0.0006, + "loss": 4.148930549621582, + "step": 5856 + }, + { + "epoch": 81.34934497816595, + "grad_norm": 0.013662154786288738, + "learning_rate": 0.0006, + "loss": 4.192107677459717, + "step": 5857 + }, + { + "epoch": 81.36331877729258, + "grad_norm": 0.013434980995953083, + "learning_rate": 0.0006, + "loss": 4.162052154541016, + "step": 5858 + }, + { + "epoch": 81.37729257641921, + "grad_norm": 0.01300742756575346, + "learning_rate": 0.0006, + "loss": 4.262385368347168, + "step": 5859 + }, + { + "epoch": 81.39126637554585, + "grad_norm": 0.012544351629912853, + "learning_rate": 0.0006, + "loss": 4.215371131896973, + "step": 5860 + }, + { + "epoch": 81.40524017467249, + "grad_norm": 0.013663645833730698, + "learning_rate": 0.0006, + "loss": 4.247312545776367, + "step": 5861 + }, + { + "epoch": 81.41921397379913, + "grad_norm": 0.013127296231687069, + "learning_rate": 0.0006, + "loss": 4.087685585021973, + "step": 5862 + }, + { + "epoch": 81.43318777292576, + "grad_norm": 0.014081938192248344, + "learning_rate": 0.0006, + "loss": 4.272648811340332, + "step": 5863 + }, + { + "epoch": 81.44716157205241, + "grad_norm": 0.015470641665160656, + "learning_rate": 0.0006, + "loss": 4.278578758239746, + "step": 5864 + }, + { + "epoch": 81.46113537117904, + "grad_norm": 0.016524532809853554, + "learning_rate": 0.0006, + "loss": 4.118268013000488, + "step": 5865 + }, + { + "epoch": 81.47510917030567, + "grad_norm": 0.016242722049355507, + "learning_rate": 0.0006, + "loss": 4.182405471801758, + "step": 5866 + }, + { + "epoch": 81.48908296943232, + "grad_norm": 0.01557017583400011, + "learning_rate": 0.0006, + "loss": 4.229862213134766, + "step": 5867 + }, + { + "epoch": 81.50305676855895, + "grad_norm": 0.01562870666384697, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 5868 + }, + { + "epoch": 81.5170305676856, + "grad_norm": 0.018606074154376984, + "learning_rate": 0.0006, + "loss": 4.148681640625, + "step": 5869 + }, + { + "epoch": 81.53100436681223, + "grad_norm": 0.017740730196237564, + "learning_rate": 0.0006, + "loss": 4.321932792663574, + "step": 5870 + }, + { + "epoch": 81.54497816593886, + "grad_norm": 0.014517677947878838, + "learning_rate": 0.0006, + "loss": 4.22862434387207, + "step": 5871 + }, + { + "epoch": 81.5589519650655, + "grad_norm": 0.01389714702963829, + "learning_rate": 0.0006, + "loss": 4.198000907897949, + "step": 5872 + }, + { + "epoch": 81.57292576419214, + "grad_norm": 0.015348682180047035, + "learning_rate": 0.0006, + "loss": 4.293485164642334, + "step": 5873 + }, + { + "epoch": 81.58689956331878, + "grad_norm": 0.015965791419148445, + "learning_rate": 0.0006, + "loss": 4.23481559753418, + "step": 5874 + }, + { + "epoch": 81.60087336244541, + "grad_norm": 0.015662726014852524, + "learning_rate": 0.0006, + "loss": 4.158148765563965, + "step": 5875 + }, + { + "epoch": 81.61484716157206, + "grad_norm": 0.014402080327272415, + "learning_rate": 0.0006, + "loss": 4.132723808288574, + "step": 5876 + }, + { + "epoch": 81.62882096069869, + "grad_norm": 0.016846125945448875, + "learning_rate": 0.0006, + "loss": 4.1654462814331055, + "step": 5877 + }, + { + "epoch": 81.64279475982532, + "grad_norm": 0.0196097232401371, + "learning_rate": 0.0006, + "loss": 4.228450298309326, + "step": 5878 + }, + { + "epoch": 81.65676855895197, + "grad_norm": 0.016005294397473335, + "learning_rate": 0.0006, + "loss": 4.230222225189209, + "step": 5879 + }, + { + "epoch": 81.6707423580786, + "grad_norm": 0.014685059897601604, + "learning_rate": 0.0006, + "loss": 4.1367387771606445, + "step": 5880 + }, + { + "epoch": 81.68471615720524, + "grad_norm": 0.01647144928574562, + "learning_rate": 0.0006, + "loss": 4.096675395965576, + "step": 5881 + }, + { + "epoch": 81.69868995633188, + "grad_norm": 0.015978610143065453, + "learning_rate": 0.0006, + "loss": 4.232292175292969, + "step": 5882 + }, + { + "epoch": 81.71266375545852, + "grad_norm": 0.01510288193821907, + "learning_rate": 0.0006, + "loss": 4.274591445922852, + "step": 5883 + }, + { + "epoch": 81.72663755458515, + "grad_norm": 0.014942476525902748, + "learning_rate": 0.0006, + "loss": 4.161482810974121, + "step": 5884 + }, + { + "epoch": 81.74061135371178, + "grad_norm": 0.014595299027860165, + "learning_rate": 0.0006, + "loss": 4.204291343688965, + "step": 5885 + }, + { + "epoch": 81.75458515283843, + "grad_norm": 0.014952429570257664, + "learning_rate": 0.0006, + "loss": 4.242499351501465, + "step": 5886 + }, + { + "epoch": 81.76855895196506, + "grad_norm": 0.014761554077267647, + "learning_rate": 0.0006, + "loss": 4.20393180847168, + "step": 5887 + }, + { + "epoch": 81.78253275109171, + "grad_norm": 0.014749647118151188, + "learning_rate": 0.0006, + "loss": 4.227565765380859, + "step": 5888 + }, + { + "epoch": 81.79650655021834, + "grad_norm": 0.014384493231773376, + "learning_rate": 0.0006, + "loss": 4.282228946685791, + "step": 5889 + }, + { + "epoch": 81.81048034934499, + "grad_norm": 0.01260330155491829, + "learning_rate": 0.0006, + "loss": 4.2882866859436035, + "step": 5890 + }, + { + "epoch": 81.82445414847162, + "grad_norm": 0.014055311679840088, + "learning_rate": 0.0006, + "loss": 4.22351598739624, + "step": 5891 + }, + { + "epoch": 81.83842794759825, + "grad_norm": 0.013674065470695496, + "learning_rate": 0.0006, + "loss": 4.240716934204102, + "step": 5892 + }, + { + "epoch": 81.8524017467249, + "grad_norm": 0.01225196197628975, + "learning_rate": 0.0006, + "loss": 4.2958598136901855, + "step": 5893 + }, + { + "epoch": 81.86637554585153, + "grad_norm": 0.013538680970668793, + "learning_rate": 0.0006, + "loss": 4.24342679977417, + "step": 5894 + }, + { + "epoch": 81.88034934497817, + "grad_norm": 0.013151394203305244, + "learning_rate": 0.0006, + "loss": 4.081999778747559, + "step": 5895 + }, + { + "epoch": 81.8943231441048, + "grad_norm": 0.014260678552091122, + "learning_rate": 0.0006, + "loss": 4.156423568725586, + "step": 5896 + }, + { + "epoch": 81.90829694323143, + "grad_norm": 0.01329776830971241, + "learning_rate": 0.0006, + "loss": 4.1713995933532715, + "step": 5897 + }, + { + "epoch": 81.92227074235808, + "grad_norm": 0.014623276889324188, + "learning_rate": 0.0006, + "loss": 4.244891166687012, + "step": 5898 + }, + { + "epoch": 81.93624454148471, + "grad_norm": 0.015760324895381927, + "learning_rate": 0.0006, + "loss": 4.076528549194336, + "step": 5899 + }, + { + "epoch": 81.95021834061136, + "grad_norm": 0.014331766404211521, + "learning_rate": 0.0006, + "loss": 4.283037185668945, + "step": 5900 + }, + { + "epoch": 81.96419213973799, + "grad_norm": 0.012383842840790749, + "learning_rate": 0.0006, + "loss": 4.117308616638184, + "step": 5901 + }, + { + "epoch": 81.97816593886463, + "grad_norm": 0.01250660140067339, + "learning_rate": 0.0006, + "loss": 4.241250991821289, + "step": 5902 + }, + { + "epoch": 81.99213973799127, + "grad_norm": 0.012276900932192802, + "learning_rate": 0.0006, + "loss": 4.141091823577881, + "step": 5903 + }, + { + "epoch": 82.0, + "grad_norm": 0.01588737778365612, + "learning_rate": 0.0006, + "loss": 4.065532684326172, + "step": 5904 + }, + { + "epoch": 82.0, + "eval_loss": 4.612304210662842, + "eval_runtime": 57.5411, + "eval_samples_per_second": 42.439, + "eval_steps_per_second": 1.338, + "step": 5904 + }, + { + "epoch": 82.01397379912663, + "grad_norm": 0.017433544620871544, + "learning_rate": 0.0006, + "loss": 4.202497482299805, + "step": 5905 + }, + { + "epoch": 82.02794759825328, + "grad_norm": 0.017260415479540825, + "learning_rate": 0.0006, + "loss": 4.099949359893799, + "step": 5906 + }, + { + "epoch": 82.04192139737991, + "grad_norm": 0.01577908545732498, + "learning_rate": 0.0006, + "loss": 4.17806339263916, + "step": 5907 + }, + { + "epoch": 82.05589519650655, + "grad_norm": 0.016093818470835686, + "learning_rate": 0.0006, + "loss": 4.219491004943848, + "step": 5908 + }, + { + "epoch": 82.06986899563319, + "grad_norm": 0.015426302328705788, + "learning_rate": 0.0006, + "loss": 4.207982063293457, + "step": 5909 + }, + { + "epoch": 82.08384279475983, + "grad_norm": 0.01605442352592945, + "learning_rate": 0.0006, + "loss": 4.265942573547363, + "step": 5910 + }, + { + "epoch": 82.09781659388646, + "grad_norm": 0.01787225715816021, + "learning_rate": 0.0006, + "loss": 4.151535511016846, + "step": 5911 + }, + { + "epoch": 82.1117903930131, + "grad_norm": 0.01750164106488228, + "learning_rate": 0.0006, + "loss": 4.133655071258545, + "step": 5912 + }, + { + "epoch": 82.12576419213974, + "grad_norm": 0.01602957211434841, + "learning_rate": 0.0006, + "loss": 4.176972389221191, + "step": 5913 + }, + { + "epoch": 82.13973799126637, + "grad_norm": 0.01553812064230442, + "learning_rate": 0.0006, + "loss": 4.125692367553711, + "step": 5914 + }, + { + "epoch": 82.15371179039302, + "grad_norm": 0.013978508301079273, + "learning_rate": 0.0006, + "loss": 4.207271575927734, + "step": 5915 + }, + { + "epoch": 82.16768558951965, + "grad_norm": 0.013561118394136429, + "learning_rate": 0.0006, + "loss": 4.238930702209473, + "step": 5916 + }, + { + "epoch": 82.18165938864628, + "grad_norm": 0.013285507448017597, + "learning_rate": 0.0006, + "loss": 4.129464626312256, + "step": 5917 + }, + { + "epoch": 82.19563318777293, + "grad_norm": 0.013903839513659477, + "learning_rate": 0.0006, + "loss": 4.214109420776367, + "step": 5918 + }, + { + "epoch": 82.20960698689956, + "grad_norm": 0.014426837675273418, + "learning_rate": 0.0006, + "loss": 4.143196105957031, + "step": 5919 + }, + { + "epoch": 82.2235807860262, + "grad_norm": 0.015347165986895561, + "learning_rate": 0.0006, + "loss": 4.235439300537109, + "step": 5920 + }, + { + "epoch": 82.23755458515284, + "grad_norm": 0.014266039244830608, + "learning_rate": 0.0006, + "loss": 3.991617202758789, + "step": 5921 + }, + { + "epoch": 82.25152838427948, + "grad_norm": 0.013101593591272831, + "learning_rate": 0.0006, + "loss": 4.266017913818359, + "step": 5922 + }, + { + "epoch": 82.26550218340611, + "grad_norm": 0.013741881586611271, + "learning_rate": 0.0006, + "loss": 4.224879741668701, + "step": 5923 + }, + { + "epoch": 82.27947598253274, + "grad_norm": 0.014206945896148682, + "learning_rate": 0.0006, + "loss": 4.102272987365723, + "step": 5924 + }, + { + "epoch": 82.29344978165939, + "grad_norm": 0.013798723928630352, + "learning_rate": 0.0006, + "loss": 4.30418586730957, + "step": 5925 + }, + { + "epoch": 82.30742358078602, + "grad_norm": 0.015470122918486595, + "learning_rate": 0.0006, + "loss": 4.2513322830200195, + "step": 5926 + }, + { + "epoch": 82.32139737991267, + "grad_norm": 0.01655442826449871, + "learning_rate": 0.0006, + "loss": 4.223002910614014, + "step": 5927 + }, + { + "epoch": 82.3353711790393, + "grad_norm": 0.017170004546642303, + "learning_rate": 0.0006, + "loss": 4.253904819488525, + "step": 5928 + }, + { + "epoch": 82.34934497816595, + "grad_norm": 0.017595812678337097, + "learning_rate": 0.0006, + "loss": 4.184927940368652, + "step": 5929 + }, + { + "epoch": 82.36331877729258, + "grad_norm": 0.015073256567120552, + "learning_rate": 0.0006, + "loss": 4.202239036560059, + "step": 5930 + }, + { + "epoch": 82.37729257641921, + "grad_norm": 0.014310941100120544, + "learning_rate": 0.0006, + "loss": 4.124934673309326, + "step": 5931 + }, + { + "epoch": 82.39126637554585, + "grad_norm": 0.013671821914613247, + "learning_rate": 0.0006, + "loss": 4.138919353485107, + "step": 5932 + }, + { + "epoch": 82.40524017467249, + "grad_norm": 0.014103314839303493, + "learning_rate": 0.0006, + "loss": 4.211289405822754, + "step": 5933 + }, + { + "epoch": 82.41921397379913, + "grad_norm": 0.014272668398916721, + "learning_rate": 0.0006, + "loss": 4.06290340423584, + "step": 5934 + }, + { + "epoch": 82.43318777292576, + "grad_norm": 0.01387016475200653, + "learning_rate": 0.0006, + "loss": 4.101334571838379, + "step": 5935 + }, + { + "epoch": 82.44716157205241, + "grad_norm": 0.013318683952093124, + "learning_rate": 0.0006, + "loss": 4.206409931182861, + "step": 5936 + }, + { + "epoch": 82.46113537117904, + "grad_norm": 0.013781268149614334, + "learning_rate": 0.0006, + "loss": 4.189474105834961, + "step": 5937 + }, + { + "epoch": 82.47510917030567, + "grad_norm": 0.013195851817727089, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 5938 + }, + { + "epoch": 82.48908296943232, + "grad_norm": 0.014358163811266422, + "learning_rate": 0.0006, + "loss": 4.203262805938721, + "step": 5939 + }, + { + "epoch": 82.50305676855895, + "grad_norm": 0.01403636671602726, + "learning_rate": 0.0006, + "loss": 4.079258441925049, + "step": 5940 + }, + { + "epoch": 82.5170305676856, + "grad_norm": 0.013272535055875778, + "learning_rate": 0.0006, + "loss": 4.19197940826416, + "step": 5941 + }, + { + "epoch": 82.53100436681223, + "grad_norm": 0.013787264935672283, + "learning_rate": 0.0006, + "loss": 4.12995719909668, + "step": 5942 + }, + { + "epoch": 82.54497816593886, + "grad_norm": 0.015020057559013367, + "learning_rate": 0.0006, + "loss": 4.148890495300293, + "step": 5943 + }, + { + "epoch": 82.5589519650655, + "grad_norm": 0.014199157245457172, + "learning_rate": 0.0006, + "loss": 4.201516151428223, + "step": 5944 + }, + { + "epoch": 82.57292576419214, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.240894317626953, + "step": 5945 + }, + { + "epoch": 82.58689956331878, + "grad_norm": 0.014767560176551342, + "learning_rate": 0.0006, + "loss": 4.173320770263672, + "step": 5946 + }, + { + "epoch": 82.60087336244541, + "grad_norm": 0.01612078957259655, + "learning_rate": 0.0006, + "loss": 4.266448020935059, + "step": 5947 + }, + { + "epoch": 82.61484716157206, + "grad_norm": 0.016706543043255806, + "learning_rate": 0.0006, + "loss": 4.229398250579834, + "step": 5948 + }, + { + "epoch": 82.62882096069869, + "grad_norm": 0.015630796551704407, + "learning_rate": 0.0006, + "loss": 4.181282997131348, + "step": 5949 + }, + { + "epoch": 82.64279475982532, + "grad_norm": 0.013103844597935677, + "learning_rate": 0.0006, + "loss": 4.309617042541504, + "step": 5950 + }, + { + "epoch": 82.65676855895197, + "grad_norm": 0.016136541962623596, + "learning_rate": 0.0006, + "loss": 4.224159240722656, + "step": 5951 + }, + { + "epoch": 82.6707423580786, + "grad_norm": 0.016160814091563225, + "learning_rate": 0.0006, + "loss": 4.267970085144043, + "step": 5952 + }, + { + "epoch": 82.68471615720524, + "grad_norm": 0.01672680303454399, + "learning_rate": 0.0006, + "loss": 4.230506896972656, + "step": 5953 + }, + { + "epoch": 82.69868995633188, + "grad_norm": 0.0159586351364851, + "learning_rate": 0.0006, + "loss": 4.159542083740234, + "step": 5954 + }, + { + "epoch": 82.71266375545852, + "grad_norm": 0.01461219135671854, + "learning_rate": 0.0006, + "loss": 4.27863073348999, + "step": 5955 + }, + { + "epoch": 82.72663755458515, + "grad_norm": 0.015088613145053387, + "learning_rate": 0.0006, + "loss": 4.272418975830078, + "step": 5956 + }, + { + "epoch": 82.74061135371178, + "grad_norm": 0.01594499684870243, + "learning_rate": 0.0006, + "loss": 4.271326065063477, + "step": 5957 + }, + { + "epoch": 82.75458515283843, + "grad_norm": 0.016648447141051292, + "learning_rate": 0.0006, + "loss": 4.141478538513184, + "step": 5958 + }, + { + "epoch": 82.76855895196506, + "grad_norm": 0.01540316641330719, + "learning_rate": 0.0006, + "loss": 4.111814498901367, + "step": 5959 + }, + { + "epoch": 82.78253275109171, + "grad_norm": 0.01634364202618599, + "learning_rate": 0.0006, + "loss": 4.3113112449646, + "step": 5960 + }, + { + "epoch": 82.79650655021834, + "grad_norm": 0.014768977649509907, + "learning_rate": 0.0006, + "loss": 4.184217929840088, + "step": 5961 + }, + { + "epoch": 82.81048034934499, + "grad_norm": 0.013453755527734756, + "learning_rate": 0.0006, + "loss": 4.247729301452637, + "step": 5962 + }, + { + "epoch": 82.82445414847162, + "grad_norm": 0.014731907285749912, + "learning_rate": 0.0006, + "loss": 4.352839469909668, + "step": 5963 + }, + { + "epoch": 82.83842794759825, + "grad_norm": 0.016047345474362373, + "learning_rate": 0.0006, + "loss": 4.257283687591553, + "step": 5964 + }, + { + "epoch": 82.8524017467249, + "grad_norm": 0.017141170799732208, + "learning_rate": 0.0006, + "loss": 4.22850227355957, + "step": 5965 + }, + { + "epoch": 82.86637554585153, + "grad_norm": 0.01370396837592125, + "learning_rate": 0.0006, + "loss": 4.1169257164001465, + "step": 5966 + }, + { + "epoch": 82.88034934497817, + "grad_norm": 0.012060081586241722, + "learning_rate": 0.0006, + "loss": 4.195242404937744, + "step": 5967 + }, + { + "epoch": 82.8943231441048, + "grad_norm": 0.015528378076851368, + "learning_rate": 0.0006, + "loss": 4.294939994812012, + "step": 5968 + }, + { + "epoch": 82.90829694323143, + "grad_norm": 0.018151041120290756, + "learning_rate": 0.0006, + "loss": 4.137192726135254, + "step": 5969 + }, + { + "epoch": 82.92227074235808, + "grad_norm": 0.017895735800266266, + "learning_rate": 0.0006, + "loss": 4.174036026000977, + "step": 5970 + }, + { + "epoch": 82.93624454148471, + "grad_norm": 0.015610386617481709, + "learning_rate": 0.0006, + "loss": 4.230136871337891, + "step": 5971 + }, + { + "epoch": 82.95021834061136, + "grad_norm": 0.01339928898960352, + "learning_rate": 0.0006, + "loss": 4.199074745178223, + "step": 5972 + }, + { + "epoch": 82.96419213973799, + "grad_norm": 0.015454214997589588, + "learning_rate": 0.0006, + "loss": 4.216109752655029, + "step": 5973 + }, + { + "epoch": 82.97816593886463, + "grad_norm": 0.014978132210671902, + "learning_rate": 0.0006, + "loss": 4.131511688232422, + "step": 5974 + }, + { + "epoch": 82.99213973799127, + "grad_norm": 0.015611122362315655, + "learning_rate": 0.0006, + "loss": 4.246518135070801, + "step": 5975 + }, + { + "epoch": 83.0, + "grad_norm": 0.016682764515280724, + "learning_rate": 0.0006, + "loss": 4.110680103302002, + "step": 5976 + }, + { + "epoch": 83.0, + "eval_loss": 4.641707897186279, + "eval_runtime": 57.2799, + "eval_samples_per_second": 42.633, + "eval_steps_per_second": 1.344, + "step": 5976 + }, + { + "epoch": 83.01397379912663, + "grad_norm": 0.015007982030510902, + "learning_rate": 0.0006, + "loss": 4.100912094116211, + "step": 5977 + }, + { + "epoch": 83.02794759825328, + "grad_norm": 0.017609436064958572, + "learning_rate": 0.0006, + "loss": 4.233399868011475, + "step": 5978 + }, + { + "epoch": 83.04192139737991, + "grad_norm": 0.019817272201180458, + "learning_rate": 0.0006, + "loss": 4.191173076629639, + "step": 5979 + }, + { + "epoch": 83.05589519650655, + "grad_norm": 0.017752140760421753, + "learning_rate": 0.0006, + "loss": 4.044437885284424, + "step": 5980 + }, + { + "epoch": 83.06986899563319, + "grad_norm": 0.015482760034501553, + "learning_rate": 0.0006, + "loss": 4.319931983947754, + "step": 5981 + }, + { + "epoch": 83.08384279475983, + "grad_norm": 0.015153988264501095, + "learning_rate": 0.0006, + "loss": 4.095277309417725, + "step": 5982 + }, + { + "epoch": 83.09781659388646, + "grad_norm": 0.017025282606482506, + "learning_rate": 0.0006, + "loss": 4.235110759735107, + "step": 5983 + }, + { + "epoch": 83.1117903930131, + "grad_norm": 0.01566586270928383, + "learning_rate": 0.0006, + "loss": 4.1934685707092285, + "step": 5984 + }, + { + "epoch": 83.12576419213974, + "grad_norm": 0.016419263556599617, + "learning_rate": 0.0006, + "loss": 4.234419822692871, + "step": 5985 + }, + { + "epoch": 83.13973799126637, + "grad_norm": 0.017132675275206566, + "learning_rate": 0.0006, + "loss": 4.221447944641113, + "step": 5986 + }, + { + "epoch": 83.15371179039302, + "grad_norm": 0.016138629987835884, + "learning_rate": 0.0006, + "loss": 4.199830055236816, + "step": 5987 + }, + { + "epoch": 83.16768558951965, + "grad_norm": 0.016210174188017845, + "learning_rate": 0.0006, + "loss": 4.081717491149902, + "step": 5988 + }, + { + "epoch": 83.18165938864628, + "grad_norm": 0.016055308282375336, + "learning_rate": 0.0006, + "loss": 4.247655391693115, + "step": 5989 + }, + { + "epoch": 83.19563318777293, + "grad_norm": 0.016132952645421028, + "learning_rate": 0.0006, + "loss": 4.262535572052002, + "step": 5990 + }, + { + "epoch": 83.20960698689956, + "grad_norm": 0.01779935695230961, + "learning_rate": 0.0006, + "loss": 4.153120040893555, + "step": 5991 + }, + { + "epoch": 83.2235807860262, + "grad_norm": 0.018360646441578865, + "learning_rate": 0.0006, + "loss": 4.232644081115723, + "step": 5992 + }, + { + "epoch": 83.23755458515284, + "grad_norm": 0.019629675894975662, + "learning_rate": 0.0006, + "loss": 3.9758734703063965, + "step": 5993 + }, + { + "epoch": 83.25152838427948, + "grad_norm": 0.017819996923208237, + "learning_rate": 0.0006, + "loss": 4.196932315826416, + "step": 5994 + }, + { + "epoch": 83.26550218340611, + "grad_norm": 0.01557681243866682, + "learning_rate": 0.0006, + "loss": 4.122654914855957, + "step": 5995 + }, + { + "epoch": 83.27947598253274, + "grad_norm": 0.017858853563666344, + "learning_rate": 0.0006, + "loss": 4.185330867767334, + "step": 5996 + }, + { + "epoch": 83.29344978165939, + "grad_norm": 0.018065953627228737, + "learning_rate": 0.0006, + "loss": 4.218471050262451, + "step": 5997 + }, + { + "epoch": 83.30742358078602, + "grad_norm": 0.017247939482331276, + "learning_rate": 0.0006, + "loss": 4.127076148986816, + "step": 5998 + }, + { + "epoch": 83.32139737991267, + "grad_norm": 0.017047952860593796, + "learning_rate": 0.0006, + "loss": 4.131356716156006, + "step": 5999 + }, + { + "epoch": 83.3353711790393, + "grad_norm": 0.015905629843473434, + "learning_rate": 0.0006, + "loss": 4.2239789962768555, + "step": 6000 + }, + { + "epoch": 83.34934497816595, + "grad_norm": 0.016522347927093506, + "learning_rate": 0.0006, + "loss": 4.054897785186768, + "step": 6001 + }, + { + "epoch": 83.36331877729258, + "grad_norm": 0.01622571051120758, + "learning_rate": 0.0006, + "loss": 4.284210681915283, + "step": 6002 + }, + { + "epoch": 83.37729257641921, + "grad_norm": 0.01640467159450054, + "learning_rate": 0.0006, + "loss": 4.268840789794922, + "step": 6003 + }, + { + "epoch": 83.39126637554585, + "grad_norm": 0.015698859468102455, + "learning_rate": 0.0006, + "loss": 4.179793357849121, + "step": 6004 + }, + { + "epoch": 83.40524017467249, + "grad_norm": 0.016694528982043266, + "learning_rate": 0.0006, + "loss": 4.132297515869141, + "step": 6005 + }, + { + "epoch": 83.41921397379913, + "grad_norm": 0.015345900319516659, + "learning_rate": 0.0006, + "loss": 4.090163230895996, + "step": 6006 + }, + { + "epoch": 83.43318777292576, + "grad_norm": 0.014722758904099464, + "learning_rate": 0.0006, + "loss": 4.179166793823242, + "step": 6007 + }, + { + "epoch": 83.44716157205241, + "grad_norm": 0.014884887263178825, + "learning_rate": 0.0006, + "loss": 4.137659072875977, + "step": 6008 + }, + { + "epoch": 83.46113537117904, + "grad_norm": 0.014041902497410774, + "learning_rate": 0.0006, + "loss": 4.1202073097229, + "step": 6009 + }, + { + "epoch": 83.47510917030567, + "grad_norm": 0.012799333781003952, + "learning_rate": 0.0006, + "loss": 4.092607498168945, + "step": 6010 + }, + { + "epoch": 83.48908296943232, + "grad_norm": 0.013272707350552082, + "learning_rate": 0.0006, + "loss": 4.132715225219727, + "step": 6011 + }, + { + "epoch": 83.50305676855895, + "grad_norm": 0.012934371829032898, + "learning_rate": 0.0006, + "loss": 4.230278968811035, + "step": 6012 + }, + { + "epoch": 83.5170305676856, + "grad_norm": 0.0120006762444973, + "learning_rate": 0.0006, + "loss": 4.140068531036377, + "step": 6013 + }, + { + "epoch": 83.53100436681223, + "grad_norm": 0.013616700656712055, + "learning_rate": 0.0006, + "loss": 4.20107364654541, + "step": 6014 + }, + { + "epoch": 83.54497816593886, + "grad_norm": 0.013957851566374302, + "learning_rate": 0.0006, + "loss": 4.124945163726807, + "step": 6015 + }, + { + "epoch": 83.5589519650655, + "grad_norm": 0.012762055732309818, + "learning_rate": 0.0006, + "loss": 4.022052764892578, + "step": 6016 + }, + { + "epoch": 83.57292576419214, + "grad_norm": 0.013080576434731483, + "learning_rate": 0.0006, + "loss": 4.103252410888672, + "step": 6017 + }, + { + "epoch": 83.58689956331878, + "grad_norm": 0.013957500457763672, + "learning_rate": 0.0006, + "loss": 4.235089302062988, + "step": 6018 + }, + { + "epoch": 83.60087336244541, + "grad_norm": 0.014707164838910103, + "learning_rate": 0.0006, + "loss": 4.191570281982422, + "step": 6019 + }, + { + "epoch": 83.61484716157206, + "grad_norm": 0.012807869352400303, + "learning_rate": 0.0006, + "loss": 4.294543266296387, + "step": 6020 + }, + { + "epoch": 83.62882096069869, + "grad_norm": 0.012027994729578495, + "learning_rate": 0.0006, + "loss": 4.13849401473999, + "step": 6021 + }, + { + "epoch": 83.64279475982532, + "grad_norm": 0.012147724628448486, + "learning_rate": 0.0006, + "loss": 4.127481937408447, + "step": 6022 + }, + { + "epoch": 83.65676855895197, + "grad_norm": 0.013241392560303211, + "learning_rate": 0.0006, + "loss": 4.218972682952881, + "step": 6023 + }, + { + "epoch": 83.6707423580786, + "grad_norm": 0.014071648940443993, + "learning_rate": 0.0006, + "loss": 4.096724033355713, + "step": 6024 + }, + { + "epoch": 83.68471615720524, + "grad_norm": 0.015212797559797764, + "learning_rate": 0.0006, + "loss": 4.27097225189209, + "step": 6025 + }, + { + "epoch": 83.69868995633188, + "grad_norm": 0.017647385597229004, + "learning_rate": 0.0006, + "loss": 4.209835529327393, + "step": 6026 + }, + { + "epoch": 83.71266375545852, + "grad_norm": 0.01709647662937641, + "learning_rate": 0.0006, + "loss": 4.042106628417969, + "step": 6027 + }, + { + "epoch": 83.72663755458515, + "grad_norm": 0.014530927874147892, + "learning_rate": 0.0006, + "loss": 4.181731700897217, + "step": 6028 + }, + { + "epoch": 83.74061135371178, + "grad_norm": 0.015356711111962795, + "learning_rate": 0.0006, + "loss": 4.314183235168457, + "step": 6029 + }, + { + "epoch": 83.75458515283843, + "grad_norm": 0.015500029549002647, + "learning_rate": 0.0006, + "loss": 4.06776762008667, + "step": 6030 + }, + { + "epoch": 83.76855895196506, + "grad_norm": 0.015351332724094391, + "learning_rate": 0.0006, + "loss": 4.185141563415527, + "step": 6031 + }, + { + "epoch": 83.78253275109171, + "grad_norm": 0.016678154468536377, + "learning_rate": 0.0006, + "loss": 4.179529666900635, + "step": 6032 + }, + { + "epoch": 83.79650655021834, + "grad_norm": 0.0156265702098608, + "learning_rate": 0.0006, + "loss": 4.238584041595459, + "step": 6033 + }, + { + "epoch": 83.81048034934499, + "grad_norm": 0.016418596729636192, + "learning_rate": 0.0006, + "loss": 4.1593523025512695, + "step": 6034 + }, + { + "epoch": 83.82445414847162, + "grad_norm": 0.016113586723804474, + "learning_rate": 0.0006, + "loss": 4.157470226287842, + "step": 6035 + }, + { + "epoch": 83.83842794759825, + "grad_norm": 0.014434042386710644, + "learning_rate": 0.0006, + "loss": 4.196226119995117, + "step": 6036 + }, + { + "epoch": 83.8524017467249, + "grad_norm": 0.012647042982280254, + "learning_rate": 0.0006, + "loss": 4.130589485168457, + "step": 6037 + }, + { + "epoch": 83.86637554585153, + "grad_norm": 0.014412877149879932, + "learning_rate": 0.0006, + "loss": 4.154845237731934, + "step": 6038 + }, + { + "epoch": 83.88034934497817, + "grad_norm": 0.014222027733922005, + "learning_rate": 0.0006, + "loss": 4.189724922180176, + "step": 6039 + }, + { + "epoch": 83.8943231441048, + "grad_norm": 0.013084255158901215, + "learning_rate": 0.0006, + "loss": 4.303338527679443, + "step": 6040 + }, + { + "epoch": 83.90829694323143, + "grad_norm": 0.012659232132136822, + "learning_rate": 0.0006, + "loss": 4.08192253112793, + "step": 6041 + }, + { + "epoch": 83.92227074235808, + "grad_norm": 0.01246276218444109, + "learning_rate": 0.0006, + "loss": 4.283324718475342, + "step": 6042 + }, + { + "epoch": 83.93624454148471, + "grad_norm": 0.01340118795633316, + "learning_rate": 0.0006, + "loss": 4.117908477783203, + "step": 6043 + }, + { + "epoch": 83.95021834061136, + "grad_norm": 0.014496722258627415, + "learning_rate": 0.0006, + "loss": 4.110318660736084, + "step": 6044 + }, + { + "epoch": 83.96419213973799, + "grad_norm": 0.013348298147320747, + "learning_rate": 0.0006, + "loss": 4.185139179229736, + "step": 6045 + }, + { + "epoch": 83.97816593886463, + "grad_norm": 0.013294277712702751, + "learning_rate": 0.0006, + "loss": 4.271645545959473, + "step": 6046 + }, + { + "epoch": 83.99213973799127, + "grad_norm": 0.014431594870984554, + "learning_rate": 0.0006, + "loss": 4.161383152008057, + "step": 6047 + }, + { + "epoch": 84.0, + "grad_norm": 0.016268927603960037, + "learning_rate": 0.0006, + "loss": 4.178324222564697, + "step": 6048 + }, + { + "epoch": 84.0, + "eval_loss": 4.6307549476623535, + "eval_runtime": 56.8521, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 6048 + }, + { + "epoch": 84.01397379912663, + "grad_norm": 0.014864908531308174, + "learning_rate": 0.0006, + "loss": 4.121300220489502, + "step": 6049 + }, + { + "epoch": 84.02794759825328, + "grad_norm": 0.014863367192447186, + "learning_rate": 0.0006, + "loss": 4.1650543212890625, + "step": 6050 + }, + { + "epoch": 84.04192139737991, + "grad_norm": 0.014379646629095078, + "learning_rate": 0.0006, + "loss": 4.0201239585876465, + "step": 6051 + }, + { + "epoch": 84.05589519650655, + "grad_norm": 0.014871201477944851, + "learning_rate": 0.0006, + "loss": 4.0820698738098145, + "step": 6052 + }, + { + "epoch": 84.06986899563319, + "grad_norm": 0.013085179962217808, + "learning_rate": 0.0006, + "loss": 4.223984718322754, + "step": 6053 + }, + { + "epoch": 84.08384279475983, + "grad_norm": 0.012485949322581291, + "learning_rate": 0.0006, + "loss": 4.182206153869629, + "step": 6054 + }, + { + "epoch": 84.09781659388646, + "grad_norm": 0.013653291389346123, + "learning_rate": 0.0006, + "loss": 4.142557144165039, + "step": 6055 + }, + { + "epoch": 84.1117903930131, + "grad_norm": 0.013930532149970531, + "learning_rate": 0.0006, + "loss": 4.13972282409668, + "step": 6056 + }, + { + "epoch": 84.12576419213974, + "grad_norm": 0.01297379657626152, + "learning_rate": 0.0006, + "loss": 4.144664764404297, + "step": 6057 + }, + { + "epoch": 84.13973799126637, + "grad_norm": 0.013210492208600044, + "learning_rate": 0.0006, + "loss": 4.253764629364014, + "step": 6058 + }, + { + "epoch": 84.15371179039302, + "grad_norm": 0.014650176279246807, + "learning_rate": 0.0006, + "loss": 4.246866226196289, + "step": 6059 + }, + { + "epoch": 84.16768558951965, + "grad_norm": 0.01482292078435421, + "learning_rate": 0.0006, + "loss": 4.140187740325928, + "step": 6060 + }, + { + "epoch": 84.18165938864628, + "grad_norm": 0.01592552848160267, + "learning_rate": 0.0006, + "loss": 4.170705318450928, + "step": 6061 + }, + { + "epoch": 84.19563318777293, + "grad_norm": 0.017152156680822372, + "learning_rate": 0.0006, + "loss": 4.126092433929443, + "step": 6062 + }, + { + "epoch": 84.20960698689956, + "grad_norm": 0.015880905091762543, + "learning_rate": 0.0006, + "loss": 4.161938190460205, + "step": 6063 + }, + { + "epoch": 84.2235807860262, + "grad_norm": 0.015451471321284771, + "learning_rate": 0.0006, + "loss": 4.263461112976074, + "step": 6064 + }, + { + "epoch": 84.23755458515284, + "grad_norm": 0.014365996234118938, + "learning_rate": 0.0006, + "loss": 4.14488410949707, + "step": 6065 + }, + { + "epoch": 84.25152838427948, + "grad_norm": 0.016364749521017075, + "learning_rate": 0.0006, + "loss": 4.112489700317383, + "step": 6066 + }, + { + "epoch": 84.26550218340611, + "grad_norm": 0.01791754551231861, + "learning_rate": 0.0006, + "loss": 4.133513450622559, + "step": 6067 + }, + { + "epoch": 84.27947598253274, + "grad_norm": 0.016609208658337593, + "learning_rate": 0.0006, + "loss": 4.201829433441162, + "step": 6068 + }, + { + "epoch": 84.29344978165939, + "grad_norm": 0.017587820068001747, + "learning_rate": 0.0006, + "loss": 4.203151702880859, + "step": 6069 + }, + { + "epoch": 84.30742358078602, + "grad_norm": 0.017047742381691933, + "learning_rate": 0.0006, + "loss": 4.077008247375488, + "step": 6070 + }, + { + "epoch": 84.32139737991267, + "grad_norm": 0.01524646207690239, + "learning_rate": 0.0006, + "loss": 4.198038101196289, + "step": 6071 + }, + { + "epoch": 84.3353711790393, + "grad_norm": 0.014196421951055527, + "learning_rate": 0.0006, + "loss": 4.164245128631592, + "step": 6072 + }, + { + "epoch": 84.34934497816595, + "grad_norm": 0.016291776672005653, + "learning_rate": 0.0006, + "loss": 4.150152206420898, + "step": 6073 + }, + { + "epoch": 84.36331877729258, + "grad_norm": 0.015627920627593994, + "learning_rate": 0.0006, + "loss": 4.110006332397461, + "step": 6074 + }, + { + "epoch": 84.37729257641921, + "grad_norm": 0.013251487165689468, + "learning_rate": 0.0006, + "loss": 4.101463317871094, + "step": 6075 + }, + { + "epoch": 84.39126637554585, + "grad_norm": 0.013601605780422688, + "learning_rate": 0.0006, + "loss": 4.191105842590332, + "step": 6076 + }, + { + "epoch": 84.40524017467249, + "grad_norm": 0.014878741465508938, + "learning_rate": 0.0006, + "loss": 4.153046131134033, + "step": 6077 + }, + { + "epoch": 84.41921397379913, + "grad_norm": 0.014783897437155247, + "learning_rate": 0.0006, + "loss": 4.362000465393066, + "step": 6078 + }, + { + "epoch": 84.43318777292576, + "grad_norm": 0.014221332035958767, + "learning_rate": 0.0006, + "loss": 4.144364833831787, + "step": 6079 + }, + { + "epoch": 84.44716157205241, + "grad_norm": 0.014911034144461155, + "learning_rate": 0.0006, + "loss": 4.088857650756836, + "step": 6080 + }, + { + "epoch": 84.46113537117904, + "grad_norm": 0.014924770221114159, + "learning_rate": 0.0006, + "loss": 4.1677446365356445, + "step": 6081 + }, + { + "epoch": 84.47510917030567, + "grad_norm": 0.014394409023225307, + "learning_rate": 0.0006, + "loss": 4.087103843688965, + "step": 6082 + }, + { + "epoch": 84.48908296943232, + "grad_norm": 0.013926676474511623, + "learning_rate": 0.0006, + "loss": 3.9972851276397705, + "step": 6083 + }, + { + "epoch": 84.50305676855895, + "grad_norm": 0.013069954700767994, + "learning_rate": 0.0006, + "loss": 3.998237133026123, + "step": 6084 + }, + { + "epoch": 84.5170305676856, + "grad_norm": 0.012748011387884617, + "learning_rate": 0.0006, + "loss": 4.195858955383301, + "step": 6085 + }, + { + "epoch": 84.53100436681223, + "grad_norm": 0.012278077192604542, + "learning_rate": 0.0006, + "loss": 4.102165222167969, + "step": 6086 + }, + { + "epoch": 84.54497816593886, + "grad_norm": 0.014259163290262222, + "learning_rate": 0.0006, + "loss": 4.092750549316406, + "step": 6087 + }, + { + "epoch": 84.5589519650655, + "grad_norm": 0.014553683809936047, + "learning_rate": 0.0006, + "loss": 4.1412763595581055, + "step": 6088 + }, + { + "epoch": 84.57292576419214, + "grad_norm": 0.013605243526399136, + "learning_rate": 0.0006, + "loss": 4.192173480987549, + "step": 6089 + }, + { + "epoch": 84.58689956331878, + "grad_norm": 0.014259358868002892, + "learning_rate": 0.0006, + "loss": 4.356374740600586, + "step": 6090 + }, + { + "epoch": 84.60087336244541, + "grad_norm": 0.014166025444865227, + "learning_rate": 0.0006, + "loss": 4.165982246398926, + "step": 6091 + }, + { + "epoch": 84.61484716157206, + "grad_norm": 0.012198803946375847, + "learning_rate": 0.0006, + "loss": 4.287566184997559, + "step": 6092 + }, + { + "epoch": 84.62882096069869, + "grad_norm": 0.013794238679111004, + "learning_rate": 0.0006, + "loss": 4.199882984161377, + "step": 6093 + }, + { + "epoch": 84.64279475982532, + "grad_norm": 0.014757642522454262, + "learning_rate": 0.0006, + "loss": 4.116573333740234, + "step": 6094 + }, + { + "epoch": 84.65676855895197, + "grad_norm": 0.014591689221560955, + "learning_rate": 0.0006, + "loss": 4.240896224975586, + "step": 6095 + }, + { + "epoch": 84.6707423580786, + "grad_norm": 0.015217579901218414, + "learning_rate": 0.0006, + "loss": 4.257142066955566, + "step": 6096 + }, + { + "epoch": 84.68471615720524, + "grad_norm": 0.015371371991932392, + "learning_rate": 0.0006, + "loss": 4.181037425994873, + "step": 6097 + }, + { + "epoch": 84.69868995633188, + "grad_norm": 0.014412983320653439, + "learning_rate": 0.0006, + "loss": 4.2207794189453125, + "step": 6098 + }, + { + "epoch": 84.71266375545852, + "grad_norm": 0.013773671351373196, + "learning_rate": 0.0006, + "loss": 4.211123466491699, + "step": 6099 + }, + { + "epoch": 84.72663755458515, + "grad_norm": 0.012838046066462994, + "learning_rate": 0.0006, + "loss": 4.175459384918213, + "step": 6100 + }, + { + "epoch": 84.74061135371178, + "grad_norm": 0.014094929210841656, + "learning_rate": 0.0006, + "loss": 4.1724958419799805, + "step": 6101 + }, + { + "epoch": 84.75458515283843, + "grad_norm": 0.014573262073099613, + "learning_rate": 0.0006, + "loss": 4.15885066986084, + "step": 6102 + }, + { + "epoch": 84.76855895196506, + "grad_norm": 0.011964819394052029, + "learning_rate": 0.0006, + "loss": 3.9804892539978027, + "step": 6103 + }, + { + "epoch": 84.78253275109171, + "grad_norm": 0.014140539802610874, + "learning_rate": 0.0006, + "loss": 4.203708648681641, + "step": 6104 + }, + { + "epoch": 84.79650655021834, + "grad_norm": 0.016618940979242325, + "learning_rate": 0.0006, + "loss": 4.147684097290039, + "step": 6105 + }, + { + "epoch": 84.81048034934499, + "grad_norm": 0.015094385482370853, + "learning_rate": 0.0006, + "loss": 4.392223834991455, + "step": 6106 + }, + { + "epoch": 84.82445414847162, + "grad_norm": 0.014386717230081558, + "learning_rate": 0.0006, + "loss": 4.253332138061523, + "step": 6107 + }, + { + "epoch": 84.83842794759825, + "grad_norm": 0.013877755962312222, + "learning_rate": 0.0006, + "loss": 4.075056076049805, + "step": 6108 + }, + { + "epoch": 84.8524017467249, + "grad_norm": 0.014062155969440937, + "learning_rate": 0.0006, + "loss": 4.072048664093018, + "step": 6109 + }, + { + "epoch": 84.86637554585153, + "grad_norm": 0.014445619657635689, + "learning_rate": 0.0006, + "loss": 4.151832580566406, + "step": 6110 + }, + { + "epoch": 84.88034934497817, + "grad_norm": 0.015116363763809204, + "learning_rate": 0.0006, + "loss": 4.088900566101074, + "step": 6111 + }, + { + "epoch": 84.8943231441048, + "grad_norm": 0.014425474219024181, + "learning_rate": 0.0006, + "loss": 4.200483322143555, + "step": 6112 + }, + { + "epoch": 84.90829694323143, + "grad_norm": 0.01449573878198862, + "learning_rate": 0.0006, + "loss": 4.239742279052734, + "step": 6113 + }, + { + "epoch": 84.92227074235808, + "grad_norm": 0.015698261559009552, + "learning_rate": 0.0006, + "loss": 4.197391510009766, + "step": 6114 + }, + { + "epoch": 84.93624454148471, + "grad_norm": 0.014545733109116554, + "learning_rate": 0.0006, + "loss": 4.035508155822754, + "step": 6115 + }, + { + "epoch": 84.95021834061136, + "grad_norm": 0.012796483002603054, + "learning_rate": 0.0006, + "loss": 4.195248126983643, + "step": 6116 + }, + { + "epoch": 84.96419213973799, + "grad_norm": 0.01264485064893961, + "learning_rate": 0.0006, + "loss": 4.158118724822998, + "step": 6117 + }, + { + "epoch": 84.97816593886463, + "grad_norm": 0.013695276342332363, + "learning_rate": 0.0006, + "loss": 4.255288124084473, + "step": 6118 + }, + { + "epoch": 84.99213973799127, + "grad_norm": 0.013459564186632633, + "learning_rate": 0.0006, + "loss": 4.204586505889893, + "step": 6119 + }, + { + "epoch": 85.0, + "grad_norm": 0.014312652871012688, + "learning_rate": 0.0006, + "loss": 4.167879104614258, + "step": 6120 + }, + { + "epoch": 85.0, + "eval_loss": 4.608152389526367, + "eval_runtime": 57.1903, + "eval_samples_per_second": 42.7, + "eval_steps_per_second": 1.346, + "step": 6120 + }, + { + "epoch": 85.01397379912663, + "grad_norm": 0.013918368145823479, + "learning_rate": 0.0006, + "loss": 4.129991054534912, + "step": 6121 + }, + { + "epoch": 85.02794759825328, + "grad_norm": 0.01604384370148182, + "learning_rate": 0.0006, + "loss": 4.233926773071289, + "step": 6122 + }, + { + "epoch": 85.04192139737991, + "grad_norm": 0.016565652564167976, + "learning_rate": 0.0006, + "loss": 4.129604339599609, + "step": 6123 + }, + { + "epoch": 85.05589519650655, + "grad_norm": 0.0149751678109169, + "learning_rate": 0.0006, + "loss": 4.165839195251465, + "step": 6124 + }, + { + "epoch": 85.06986899563319, + "grad_norm": 0.013812461867928505, + "learning_rate": 0.0006, + "loss": 4.153792858123779, + "step": 6125 + }, + { + "epoch": 85.08384279475983, + "grad_norm": 0.014432745985686779, + "learning_rate": 0.0006, + "loss": 4.170895576477051, + "step": 6126 + }, + { + "epoch": 85.09781659388646, + "grad_norm": 0.015208868309855461, + "learning_rate": 0.0006, + "loss": 4.10639762878418, + "step": 6127 + }, + { + "epoch": 85.1117903930131, + "grad_norm": 0.013209272176027298, + "learning_rate": 0.0006, + "loss": 4.200364112854004, + "step": 6128 + }, + { + "epoch": 85.12576419213974, + "grad_norm": 0.014807002618908882, + "learning_rate": 0.0006, + "loss": 4.172666549682617, + "step": 6129 + }, + { + "epoch": 85.13973799126637, + "grad_norm": 0.015662945806980133, + "learning_rate": 0.0006, + "loss": 4.145219326019287, + "step": 6130 + }, + { + "epoch": 85.15371179039302, + "grad_norm": 0.014598497189581394, + "learning_rate": 0.0006, + "loss": 4.275618553161621, + "step": 6131 + }, + { + "epoch": 85.16768558951965, + "grad_norm": 0.013778576627373695, + "learning_rate": 0.0006, + "loss": 4.204716682434082, + "step": 6132 + }, + { + "epoch": 85.18165938864628, + "grad_norm": 0.015762265771627426, + "learning_rate": 0.0006, + "loss": 4.145524024963379, + "step": 6133 + }, + { + "epoch": 85.19563318777293, + "grad_norm": 0.015579195693135262, + "learning_rate": 0.0006, + "loss": 4.150625228881836, + "step": 6134 + }, + { + "epoch": 85.20960698689956, + "grad_norm": 0.016641216352581978, + "learning_rate": 0.0006, + "loss": 4.141819953918457, + "step": 6135 + }, + { + "epoch": 85.2235807860262, + "grad_norm": 0.01632271707057953, + "learning_rate": 0.0006, + "loss": 4.030620574951172, + "step": 6136 + }, + { + "epoch": 85.23755458515284, + "grad_norm": 0.015564918518066406, + "learning_rate": 0.0006, + "loss": 4.147753715515137, + "step": 6137 + }, + { + "epoch": 85.25152838427948, + "grad_norm": 0.0161129217594862, + "learning_rate": 0.0006, + "loss": 4.128524303436279, + "step": 6138 + }, + { + "epoch": 85.26550218340611, + "grad_norm": 0.017766550183296204, + "learning_rate": 0.0006, + "loss": 4.134856224060059, + "step": 6139 + }, + { + "epoch": 85.27947598253274, + "grad_norm": 0.01869054324924946, + "learning_rate": 0.0006, + "loss": 4.231565475463867, + "step": 6140 + }, + { + "epoch": 85.29344978165939, + "grad_norm": 0.01917010359466076, + "learning_rate": 0.0006, + "loss": 4.21175479888916, + "step": 6141 + }, + { + "epoch": 85.30742358078602, + "grad_norm": 0.018116485327482224, + "learning_rate": 0.0006, + "loss": 4.122644424438477, + "step": 6142 + }, + { + "epoch": 85.32139737991267, + "grad_norm": 0.015949107706546783, + "learning_rate": 0.0006, + "loss": 4.135458946228027, + "step": 6143 + }, + { + "epoch": 85.3353711790393, + "grad_norm": 0.014783683232963085, + "learning_rate": 0.0006, + "loss": 4.118994235992432, + "step": 6144 + }, + { + "epoch": 85.34934497816595, + "grad_norm": 0.015406692400574684, + "learning_rate": 0.0006, + "loss": 4.157268524169922, + "step": 6145 + }, + { + "epoch": 85.36331877729258, + "grad_norm": 0.013893205672502518, + "learning_rate": 0.0006, + "loss": 4.056166648864746, + "step": 6146 + }, + { + "epoch": 85.37729257641921, + "grad_norm": 0.014380923472344875, + "learning_rate": 0.0006, + "loss": 4.206905364990234, + "step": 6147 + }, + { + "epoch": 85.39126637554585, + "grad_norm": 0.015669258311390877, + "learning_rate": 0.0006, + "loss": 4.213881492614746, + "step": 6148 + }, + { + "epoch": 85.40524017467249, + "grad_norm": 0.01553428452461958, + "learning_rate": 0.0006, + "loss": 4.1814374923706055, + "step": 6149 + }, + { + "epoch": 85.41921397379913, + "grad_norm": 0.01697700284421444, + "learning_rate": 0.0006, + "loss": 4.083939552307129, + "step": 6150 + }, + { + "epoch": 85.43318777292576, + "grad_norm": 0.016639070585370064, + "learning_rate": 0.0006, + "loss": 4.279311180114746, + "step": 6151 + }, + { + "epoch": 85.44716157205241, + "grad_norm": 0.016285281628370285, + "learning_rate": 0.0006, + "loss": 4.244845390319824, + "step": 6152 + }, + { + "epoch": 85.46113537117904, + "grad_norm": 0.016071317717432976, + "learning_rate": 0.0006, + "loss": 4.089814186096191, + "step": 6153 + }, + { + "epoch": 85.47510917030567, + "grad_norm": 0.01640418916940689, + "learning_rate": 0.0006, + "loss": 4.168088912963867, + "step": 6154 + }, + { + "epoch": 85.48908296943232, + "grad_norm": 0.014781414531171322, + "learning_rate": 0.0006, + "loss": 4.196092128753662, + "step": 6155 + }, + { + "epoch": 85.50305676855895, + "grad_norm": 0.013393213972449303, + "learning_rate": 0.0006, + "loss": 4.193429946899414, + "step": 6156 + }, + { + "epoch": 85.5170305676856, + "grad_norm": 0.01467406190931797, + "learning_rate": 0.0006, + "loss": 4.216714859008789, + "step": 6157 + }, + { + "epoch": 85.53100436681223, + "grad_norm": 0.014950765296816826, + "learning_rate": 0.0006, + "loss": 4.093854904174805, + "step": 6158 + }, + { + "epoch": 85.54497816593886, + "grad_norm": 0.014655301347374916, + "learning_rate": 0.0006, + "loss": 4.09483528137207, + "step": 6159 + }, + { + "epoch": 85.5589519650655, + "grad_norm": 0.01425520796328783, + "learning_rate": 0.0006, + "loss": 4.090396881103516, + "step": 6160 + }, + { + "epoch": 85.57292576419214, + "grad_norm": 0.01403278298676014, + "learning_rate": 0.0006, + "loss": 4.088724136352539, + "step": 6161 + }, + { + "epoch": 85.58689956331878, + "grad_norm": 0.01364967692643404, + "learning_rate": 0.0006, + "loss": 4.095558166503906, + "step": 6162 + }, + { + "epoch": 85.60087336244541, + "grad_norm": 0.013085267506539822, + "learning_rate": 0.0006, + "loss": 4.163962364196777, + "step": 6163 + }, + { + "epoch": 85.61484716157206, + "grad_norm": 0.0136906448751688, + "learning_rate": 0.0006, + "loss": 4.197134017944336, + "step": 6164 + }, + { + "epoch": 85.62882096069869, + "grad_norm": 0.013472451828420162, + "learning_rate": 0.0006, + "loss": 4.1578369140625, + "step": 6165 + }, + { + "epoch": 85.64279475982532, + "grad_norm": 0.012329773046076298, + "learning_rate": 0.0006, + "loss": 4.231907844543457, + "step": 6166 + }, + { + "epoch": 85.65676855895197, + "grad_norm": 0.011977200396358967, + "learning_rate": 0.0006, + "loss": 4.1552958488464355, + "step": 6167 + }, + { + "epoch": 85.6707423580786, + "grad_norm": 0.013049956411123276, + "learning_rate": 0.0006, + "loss": 4.17852783203125, + "step": 6168 + }, + { + "epoch": 85.68471615720524, + "grad_norm": 0.011440741829574108, + "learning_rate": 0.0006, + "loss": 4.144646644592285, + "step": 6169 + }, + { + "epoch": 85.69868995633188, + "grad_norm": 0.012472936883568764, + "learning_rate": 0.0006, + "loss": 4.2306413650512695, + "step": 6170 + }, + { + "epoch": 85.71266375545852, + "grad_norm": 0.013015356846153736, + "learning_rate": 0.0006, + "loss": 4.193439960479736, + "step": 6171 + }, + { + "epoch": 85.72663755458515, + "grad_norm": 0.013832319527864456, + "learning_rate": 0.0006, + "loss": 4.18073034286499, + "step": 6172 + }, + { + "epoch": 85.74061135371178, + "grad_norm": 0.014613982290029526, + "learning_rate": 0.0006, + "loss": 4.223531246185303, + "step": 6173 + }, + { + "epoch": 85.75458515283843, + "grad_norm": 0.014209083281457424, + "learning_rate": 0.0006, + "loss": 4.142666816711426, + "step": 6174 + }, + { + "epoch": 85.76855895196506, + "grad_norm": 0.013396669179201126, + "learning_rate": 0.0006, + "loss": 4.185786247253418, + "step": 6175 + }, + { + "epoch": 85.78253275109171, + "grad_norm": 0.01364929135888815, + "learning_rate": 0.0006, + "loss": 4.125226974487305, + "step": 6176 + }, + { + "epoch": 85.79650655021834, + "grad_norm": 0.012561003677546978, + "learning_rate": 0.0006, + "loss": 4.084519386291504, + "step": 6177 + }, + { + "epoch": 85.81048034934499, + "grad_norm": 0.013508481904864311, + "learning_rate": 0.0006, + "loss": 4.193173885345459, + "step": 6178 + }, + { + "epoch": 85.82445414847162, + "grad_norm": 0.014154613018035889, + "learning_rate": 0.0006, + "loss": 4.257272720336914, + "step": 6179 + }, + { + "epoch": 85.83842794759825, + "grad_norm": 0.012478531338274479, + "learning_rate": 0.0006, + "loss": 4.161380767822266, + "step": 6180 + }, + { + "epoch": 85.8524017467249, + "grad_norm": 0.013468451797962189, + "learning_rate": 0.0006, + "loss": 4.121395111083984, + "step": 6181 + }, + { + "epoch": 85.86637554585153, + "grad_norm": 0.015155194327235222, + "learning_rate": 0.0006, + "loss": 4.242815971374512, + "step": 6182 + }, + { + "epoch": 85.88034934497817, + "grad_norm": 0.016223272308707237, + "learning_rate": 0.0006, + "loss": 4.176027297973633, + "step": 6183 + }, + { + "epoch": 85.8943231441048, + "grad_norm": 0.013741295784711838, + "learning_rate": 0.0006, + "loss": 4.131257057189941, + "step": 6184 + }, + { + "epoch": 85.90829694323143, + "grad_norm": 0.014982067979872227, + "learning_rate": 0.0006, + "loss": 4.237896919250488, + "step": 6185 + }, + { + "epoch": 85.92227074235808, + "grad_norm": 0.018197322264313698, + "learning_rate": 0.0006, + "loss": 4.154962539672852, + "step": 6186 + }, + { + "epoch": 85.93624454148471, + "grad_norm": 0.01971200481057167, + "learning_rate": 0.0006, + "loss": 4.16819429397583, + "step": 6187 + }, + { + "epoch": 85.95021834061136, + "grad_norm": 0.018008720129728317, + "learning_rate": 0.0006, + "loss": 4.146164417266846, + "step": 6188 + }, + { + "epoch": 85.96419213973799, + "grad_norm": 0.015859264880418777, + "learning_rate": 0.0006, + "loss": 4.208099365234375, + "step": 6189 + }, + { + "epoch": 85.97816593886463, + "grad_norm": 0.013846252113580704, + "learning_rate": 0.0006, + "loss": 4.1881184577941895, + "step": 6190 + }, + { + "epoch": 85.99213973799127, + "grad_norm": 0.018641939386725426, + "learning_rate": 0.0006, + "loss": 4.127632141113281, + "step": 6191 + }, + { + "epoch": 86.0, + "grad_norm": 0.02385227382183075, + "learning_rate": 0.0006, + "loss": 4.281900405883789, + "step": 6192 + }, + { + "epoch": 86.0, + "eval_loss": 4.625458717346191, + "eval_runtime": 57.0634, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6192 + }, + { + "epoch": 86.01397379912663, + "grad_norm": 0.019032124429941177, + "learning_rate": 0.0006, + "loss": 4.2163190841674805, + "step": 6193 + }, + { + "epoch": 86.02794759825328, + "grad_norm": 0.018049925565719604, + "learning_rate": 0.0006, + "loss": 4.1534037590026855, + "step": 6194 + }, + { + "epoch": 86.04192139737991, + "grad_norm": 0.019116360694169998, + "learning_rate": 0.0006, + "loss": 4.212411403656006, + "step": 6195 + }, + { + "epoch": 86.05589519650655, + "grad_norm": 0.018967336043715477, + "learning_rate": 0.0006, + "loss": 4.190918922424316, + "step": 6196 + }, + { + "epoch": 86.06986899563319, + "grad_norm": 0.01945611648261547, + "learning_rate": 0.0006, + "loss": 4.18459415435791, + "step": 6197 + }, + { + "epoch": 86.08384279475983, + "grad_norm": 0.018104076385498047, + "learning_rate": 0.0006, + "loss": 4.0942816734313965, + "step": 6198 + }, + { + "epoch": 86.09781659388646, + "grad_norm": 0.01815992221236229, + "learning_rate": 0.0006, + "loss": 4.011547565460205, + "step": 6199 + }, + { + "epoch": 86.1117903930131, + "grad_norm": 0.017574431374669075, + "learning_rate": 0.0006, + "loss": 4.051593780517578, + "step": 6200 + }, + { + "epoch": 86.12576419213974, + "grad_norm": 0.01516508124768734, + "learning_rate": 0.0006, + "loss": 4.239673614501953, + "step": 6201 + }, + { + "epoch": 86.13973799126637, + "grad_norm": 0.015689702704548836, + "learning_rate": 0.0006, + "loss": 4.1085734367370605, + "step": 6202 + }, + { + "epoch": 86.15371179039302, + "grad_norm": 0.01596900261938572, + "learning_rate": 0.0006, + "loss": 4.1509599685668945, + "step": 6203 + }, + { + "epoch": 86.16768558951965, + "grad_norm": 0.01550096832215786, + "learning_rate": 0.0006, + "loss": 4.160238265991211, + "step": 6204 + }, + { + "epoch": 86.18165938864628, + "grad_norm": 0.014601529575884342, + "learning_rate": 0.0006, + "loss": 4.029219150543213, + "step": 6205 + }, + { + "epoch": 86.19563318777293, + "grad_norm": 0.014006740413606167, + "learning_rate": 0.0006, + "loss": 4.117501258850098, + "step": 6206 + }, + { + "epoch": 86.20960698689956, + "grad_norm": 0.013191957026720047, + "learning_rate": 0.0006, + "loss": 4.252002716064453, + "step": 6207 + }, + { + "epoch": 86.2235807860262, + "grad_norm": 0.013031814247369766, + "learning_rate": 0.0006, + "loss": 4.209517478942871, + "step": 6208 + }, + { + "epoch": 86.23755458515284, + "grad_norm": 0.013605714775621891, + "learning_rate": 0.0006, + "loss": 4.176309585571289, + "step": 6209 + }, + { + "epoch": 86.25152838427948, + "grad_norm": 0.012753087095916271, + "learning_rate": 0.0006, + "loss": 4.097990036010742, + "step": 6210 + }, + { + "epoch": 86.26550218340611, + "grad_norm": 0.012940628454089165, + "learning_rate": 0.0006, + "loss": 4.085589408874512, + "step": 6211 + }, + { + "epoch": 86.27947598253274, + "grad_norm": 0.014286619611084461, + "learning_rate": 0.0006, + "loss": 4.199261665344238, + "step": 6212 + }, + { + "epoch": 86.29344978165939, + "grad_norm": 0.014643525704741478, + "learning_rate": 0.0006, + "loss": 4.23718786239624, + "step": 6213 + }, + { + "epoch": 86.30742358078602, + "grad_norm": 0.015182147733867168, + "learning_rate": 0.0006, + "loss": 4.219855308532715, + "step": 6214 + }, + { + "epoch": 86.32139737991267, + "grad_norm": 0.01530102826654911, + "learning_rate": 0.0006, + "loss": 4.190094947814941, + "step": 6215 + }, + { + "epoch": 86.3353711790393, + "grad_norm": 0.015492603182792664, + "learning_rate": 0.0006, + "loss": 4.153780460357666, + "step": 6216 + }, + { + "epoch": 86.34934497816595, + "grad_norm": 0.015339907258749008, + "learning_rate": 0.0006, + "loss": 4.182865619659424, + "step": 6217 + }, + { + "epoch": 86.36331877729258, + "grad_norm": 0.015576880425214767, + "learning_rate": 0.0006, + "loss": 4.196417331695557, + "step": 6218 + }, + { + "epoch": 86.37729257641921, + "grad_norm": 0.014805737882852554, + "learning_rate": 0.0006, + "loss": 4.265018463134766, + "step": 6219 + }, + { + "epoch": 86.39126637554585, + "grad_norm": 0.01597665622830391, + "learning_rate": 0.0006, + "loss": 4.2864837646484375, + "step": 6220 + }, + { + "epoch": 86.40524017467249, + "grad_norm": 0.01693643443286419, + "learning_rate": 0.0006, + "loss": 4.118192672729492, + "step": 6221 + }, + { + "epoch": 86.41921397379913, + "grad_norm": 0.016648951917886734, + "learning_rate": 0.0006, + "loss": 4.23712682723999, + "step": 6222 + }, + { + "epoch": 86.43318777292576, + "grad_norm": 0.014809766784310341, + "learning_rate": 0.0006, + "loss": 4.251391410827637, + "step": 6223 + }, + { + "epoch": 86.44716157205241, + "grad_norm": 0.014333373866975307, + "learning_rate": 0.0006, + "loss": 4.17243766784668, + "step": 6224 + }, + { + "epoch": 86.46113537117904, + "grad_norm": 0.013912711292505264, + "learning_rate": 0.0006, + "loss": 4.250836372375488, + "step": 6225 + }, + { + "epoch": 86.47510917030567, + "grad_norm": 0.014821341261267662, + "learning_rate": 0.0006, + "loss": 4.073995590209961, + "step": 6226 + }, + { + "epoch": 86.48908296943232, + "grad_norm": 0.016634922474622726, + "learning_rate": 0.0006, + "loss": 4.155867576599121, + "step": 6227 + }, + { + "epoch": 86.50305676855895, + "grad_norm": 0.01748695969581604, + "learning_rate": 0.0006, + "loss": 4.139344692230225, + "step": 6228 + }, + { + "epoch": 86.5170305676856, + "grad_norm": 0.01569363661110401, + "learning_rate": 0.0006, + "loss": 4.23923397064209, + "step": 6229 + }, + { + "epoch": 86.53100436681223, + "grad_norm": 0.014037121087312698, + "learning_rate": 0.0006, + "loss": 4.200334072113037, + "step": 6230 + }, + { + "epoch": 86.54497816593886, + "grad_norm": 0.01401279866695404, + "learning_rate": 0.0006, + "loss": 4.035246849060059, + "step": 6231 + }, + { + "epoch": 86.5589519650655, + "grad_norm": 0.014603731222450733, + "learning_rate": 0.0006, + "loss": 4.141385555267334, + "step": 6232 + }, + { + "epoch": 86.57292576419214, + "grad_norm": 0.015113704837858677, + "learning_rate": 0.0006, + "loss": 4.130824089050293, + "step": 6233 + }, + { + "epoch": 86.58689956331878, + "grad_norm": 0.01588696800172329, + "learning_rate": 0.0006, + "loss": 4.05495023727417, + "step": 6234 + }, + { + "epoch": 86.60087336244541, + "grad_norm": 0.01593640074133873, + "learning_rate": 0.0006, + "loss": 4.22935152053833, + "step": 6235 + }, + { + "epoch": 86.61484716157206, + "grad_norm": 0.014292445965111256, + "learning_rate": 0.0006, + "loss": 4.309839248657227, + "step": 6236 + }, + { + "epoch": 86.62882096069869, + "grad_norm": 0.015471610240638256, + "learning_rate": 0.0006, + "loss": 4.268278121948242, + "step": 6237 + }, + { + "epoch": 86.64279475982532, + "grad_norm": 0.016643544659018517, + "learning_rate": 0.0006, + "loss": 4.135109901428223, + "step": 6238 + }, + { + "epoch": 86.65676855895197, + "grad_norm": 0.014733029529452324, + "learning_rate": 0.0006, + "loss": 4.060197830200195, + "step": 6239 + }, + { + "epoch": 86.6707423580786, + "grad_norm": 0.014869216829538345, + "learning_rate": 0.0006, + "loss": 4.126917362213135, + "step": 6240 + }, + { + "epoch": 86.68471615720524, + "grad_norm": 0.01608203910291195, + "learning_rate": 0.0006, + "loss": 4.183716773986816, + "step": 6241 + }, + { + "epoch": 86.69868995633188, + "grad_norm": 0.016178175806999207, + "learning_rate": 0.0006, + "loss": 4.118875503540039, + "step": 6242 + }, + { + "epoch": 86.71266375545852, + "grad_norm": 0.014576597139239311, + "learning_rate": 0.0006, + "loss": 4.261541366577148, + "step": 6243 + }, + { + "epoch": 86.72663755458515, + "grad_norm": 0.014466420747339725, + "learning_rate": 0.0006, + "loss": 4.307284355163574, + "step": 6244 + }, + { + "epoch": 86.74061135371178, + "grad_norm": 0.014042644761502743, + "learning_rate": 0.0006, + "loss": 4.2729997634887695, + "step": 6245 + }, + { + "epoch": 86.75458515283843, + "grad_norm": 0.014589524827897549, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6246 + }, + { + "epoch": 86.76855895196506, + "grad_norm": 0.018765483051538467, + "learning_rate": 0.0006, + "loss": 4.264769554138184, + "step": 6247 + }, + { + "epoch": 86.78253275109171, + "grad_norm": 0.019397731870412827, + "learning_rate": 0.0006, + "loss": 4.183379173278809, + "step": 6248 + }, + { + "epoch": 86.79650655021834, + "grad_norm": 0.017525402829051018, + "learning_rate": 0.0006, + "loss": 4.1494245529174805, + "step": 6249 + }, + { + "epoch": 86.81048034934499, + "grad_norm": 0.016310200095176697, + "learning_rate": 0.0006, + "loss": 4.149104118347168, + "step": 6250 + }, + { + "epoch": 86.82445414847162, + "grad_norm": 0.01506423857063055, + "learning_rate": 0.0006, + "loss": 4.22091817855835, + "step": 6251 + }, + { + "epoch": 86.83842794759825, + "grad_norm": 0.014200284145772457, + "learning_rate": 0.0006, + "loss": 4.177569389343262, + "step": 6252 + }, + { + "epoch": 86.8524017467249, + "grad_norm": 0.015016552060842514, + "learning_rate": 0.0006, + "loss": 4.089859962463379, + "step": 6253 + }, + { + "epoch": 86.86637554585153, + "grad_norm": 0.016054971143603325, + "learning_rate": 0.0006, + "loss": 4.157166004180908, + "step": 6254 + }, + { + "epoch": 86.88034934497817, + "grad_norm": 0.014388521201908588, + "learning_rate": 0.0006, + "loss": 4.1437578201293945, + "step": 6255 + }, + { + "epoch": 86.8943231441048, + "grad_norm": 0.015705369412899017, + "learning_rate": 0.0006, + "loss": 4.045689582824707, + "step": 6256 + }, + { + "epoch": 86.90829694323143, + "grad_norm": 0.014192395843565464, + "learning_rate": 0.0006, + "loss": 4.098793029785156, + "step": 6257 + }, + { + "epoch": 86.92227074235808, + "grad_norm": 0.013751034624874592, + "learning_rate": 0.0006, + "loss": 3.989920139312744, + "step": 6258 + }, + { + "epoch": 86.93624454148471, + "grad_norm": 0.01582193188369274, + "learning_rate": 0.0006, + "loss": 4.283449172973633, + "step": 6259 + }, + { + "epoch": 86.95021834061136, + "grad_norm": 0.01476013008505106, + "learning_rate": 0.0006, + "loss": 4.221401691436768, + "step": 6260 + }, + { + "epoch": 86.96419213973799, + "grad_norm": 0.01346337515860796, + "learning_rate": 0.0006, + "loss": 4.05896520614624, + "step": 6261 + }, + { + "epoch": 86.97816593886463, + "grad_norm": 0.012899307534098625, + "learning_rate": 0.0006, + "loss": 4.159560203552246, + "step": 6262 + }, + { + "epoch": 86.99213973799127, + "grad_norm": 0.013150574639439583, + "learning_rate": 0.0006, + "loss": 4.217330455780029, + "step": 6263 + }, + { + "epoch": 87.0, + "grad_norm": 0.014940370805561543, + "learning_rate": 0.0006, + "loss": 4.1611833572387695, + "step": 6264 + }, + { + "epoch": 87.0, + "eval_loss": 4.666043758392334, + "eval_runtime": 56.5978, + "eval_samples_per_second": 43.147, + "eval_steps_per_second": 1.36, + "step": 6264 + }, + { + "epoch": 87.01397379912663, + "grad_norm": 0.013721180148422718, + "learning_rate": 0.0006, + "loss": 4.142454147338867, + "step": 6265 + }, + { + "epoch": 87.02794759825328, + "grad_norm": 0.014805255457758904, + "learning_rate": 0.0006, + "loss": 4.121677875518799, + "step": 6266 + }, + { + "epoch": 87.04192139737991, + "grad_norm": 0.015576514415442944, + "learning_rate": 0.0006, + "loss": 4.1188435554504395, + "step": 6267 + }, + { + "epoch": 87.05589519650655, + "grad_norm": 0.01605502888560295, + "learning_rate": 0.0006, + "loss": 4.255428314208984, + "step": 6268 + }, + { + "epoch": 87.06986899563319, + "grad_norm": 0.014933343045413494, + "learning_rate": 0.0006, + "loss": 4.122596740722656, + "step": 6269 + }, + { + "epoch": 87.08384279475983, + "grad_norm": 0.014630894176661968, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 6270 + }, + { + "epoch": 87.09781659388646, + "grad_norm": 0.014211163856089115, + "learning_rate": 0.0006, + "loss": 4.058030128479004, + "step": 6271 + }, + { + "epoch": 87.1117903930131, + "grad_norm": 0.014060168527066708, + "learning_rate": 0.0006, + "loss": 4.146586894989014, + "step": 6272 + }, + { + "epoch": 87.12576419213974, + "grad_norm": 0.014740100130438805, + "learning_rate": 0.0006, + "loss": 4.077739715576172, + "step": 6273 + }, + { + "epoch": 87.13973799126637, + "grad_norm": 0.013970664702355862, + "learning_rate": 0.0006, + "loss": 4.038252830505371, + "step": 6274 + }, + { + "epoch": 87.15371179039302, + "grad_norm": 0.014399951323866844, + "learning_rate": 0.0006, + "loss": 4.143700122833252, + "step": 6275 + }, + { + "epoch": 87.16768558951965, + "grad_norm": 0.01496410183608532, + "learning_rate": 0.0006, + "loss": 4.105609893798828, + "step": 6276 + }, + { + "epoch": 87.18165938864628, + "grad_norm": 0.014287355355918407, + "learning_rate": 0.0006, + "loss": 4.173645973205566, + "step": 6277 + }, + { + "epoch": 87.19563318777293, + "grad_norm": 0.0141292130574584, + "learning_rate": 0.0006, + "loss": 4.068511486053467, + "step": 6278 + }, + { + "epoch": 87.20960698689956, + "grad_norm": 0.01393523346632719, + "learning_rate": 0.0006, + "loss": 4.060687065124512, + "step": 6279 + }, + { + "epoch": 87.2235807860262, + "grad_norm": 0.01431021187454462, + "learning_rate": 0.0006, + "loss": 4.219209671020508, + "step": 6280 + }, + { + "epoch": 87.23755458515284, + "grad_norm": 0.014856027439236641, + "learning_rate": 0.0006, + "loss": 4.146749973297119, + "step": 6281 + }, + { + "epoch": 87.25152838427948, + "grad_norm": 0.013677539303898811, + "learning_rate": 0.0006, + "loss": 3.9859797954559326, + "step": 6282 + }, + { + "epoch": 87.26550218340611, + "grad_norm": 0.014078967273235321, + "learning_rate": 0.0006, + "loss": 4.129668235778809, + "step": 6283 + }, + { + "epoch": 87.27947598253274, + "grad_norm": 0.015564394183456898, + "learning_rate": 0.0006, + "loss": 4.0007829666137695, + "step": 6284 + }, + { + "epoch": 87.29344978165939, + "grad_norm": 0.015077749267220497, + "learning_rate": 0.0006, + "loss": 4.112264156341553, + "step": 6285 + }, + { + "epoch": 87.30742358078602, + "grad_norm": 0.015341218560934067, + "learning_rate": 0.0006, + "loss": 4.112249374389648, + "step": 6286 + }, + { + "epoch": 87.32139737991267, + "grad_norm": 0.016399366781115532, + "learning_rate": 0.0006, + "loss": 4.176196098327637, + "step": 6287 + }, + { + "epoch": 87.3353711790393, + "grad_norm": 0.01599763333797455, + "learning_rate": 0.0006, + "loss": 4.163441181182861, + "step": 6288 + }, + { + "epoch": 87.34934497816595, + "grad_norm": 0.015439776703715324, + "learning_rate": 0.0006, + "loss": 4.11902379989624, + "step": 6289 + }, + { + "epoch": 87.36331877729258, + "grad_norm": 0.015909943729639053, + "learning_rate": 0.0006, + "loss": 4.16621732711792, + "step": 6290 + }, + { + "epoch": 87.37729257641921, + "grad_norm": 0.016710078343749046, + "learning_rate": 0.0006, + "loss": 4.044960975646973, + "step": 6291 + }, + { + "epoch": 87.39126637554585, + "grad_norm": 0.015883933752775192, + "learning_rate": 0.0006, + "loss": 4.065184593200684, + "step": 6292 + }, + { + "epoch": 87.40524017467249, + "grad_norm": 0.013912747614085674, + "learning_rate": 0.0006, + "loss": 4.102450370788574, + "step": 6293 + }, + { + "epoch": 87.41921397379913, + "grad_norm": 0.013170420192182064, + "learning_rate": 0.0006, + "loss": 4.234364032745361, + "step": 6294 + }, + { + "epoch": 87.43318777292576, + "grad_norm": 0.014043302275240421, + "learning_rate": 0.0006, + "loss": 4.129565238952637, + "step": 6295 + }, + { + "epoch": 87.44716157205241, + "grad_norm": 0.015010139904916286, + "learning_rate": 0.0006, + "loss": 4.166678428649902, + "step": 6296 + }, + { + "epoch": 87.46113537117904, + "grad_norm": 0.014721784740686417, + "learning_rate": 0.0006, + "loss": 4.182618618011475, + "step": 6297 + }, + { + "epoch": 87.47510917030567, + "grad_norm": 0.015713468194007874, + "learning_rate": 0.0006, + "loss": 4.103855133056641, + "step": 6298 + }, + { + "epoch": 87.48908296943232, + "grad_norm": 0.017204247415065765, + "learning_rate": 0.0006, + "loss": 4.140937328338623, + "step": 6299 + }, + { + "epoch": 87.50305676855895, + "grad_norm": 0.016856668516993523, + "learning_rate": 0.0006, + "loss": 4.147859573364258, + "step": 6300 + }, + { + "epoch": 87.5170305676856, + "grad_norm": 0.017205018550157547, + "learning_rate": 0.0006, + "loss": 4.1667327880859375, + "step": 6301 + }, + { + "epoch": 87.53100436681223, + "grad_norm": 0.014821539632976055, + "learning_rate": 0.0006, + "loss": 4.089935302734375, + "step": 6302 + }, + { + "epoch": 87.54497816593886, + "grad_norm": 0.01446484588086605, + "learning_rate": 0.0006, + "loss": 4.243657112121582, + "step": 6303 + }, + { + "epoch": 87.5589519650655, + "grad_norm": 0.014374484308063984, + "learning_rate": 0.0006, + "loss": 4.14356803894043, + "step": 6304 + }, + { + "epoch": 87.57292576419214, + "grad_norm": 0.01420109998434782, + "learning_rate": 0.0006, + "loss": 4.1273651123046875, + "step": 6305 + }, + { + "epoch": 87.58689956331878, + "grad_norm": 0.014810388907790184, + "learning_rate": 0.0006, + "loss": 4.222290515899658, + "step": 6306 + }, + { + "epoch": 87.60087336244541, + "grad_norm": 0.015826983377337456, + "learning_rate": 0.0006, + "loss": 4.137996196746826, + "step": 6307 + }, + { + "epoch": 87.61484716157206, + "grad_norm": 0.014457454904913902, + "learning_rate": 0.0006, + "loss": 4.0780744552612305, + "step": 6308 + }, + { + "epoch": 87.62882096069869, + "grad_norm": 0.014458604156970978, + "learning_rate": 0.0006, + "loss": 4.1287922859191895, + "step": 6309 + }, + { + "epoch": 87.64279475982532, + "grad_norm": 0.015282071195542812, + "learning_rate": 0.0006, + "loss": 4.189508438110352, + "step": 6310 + }, + { + "epoch": 87.65676855895197, + "grad_norm": 0.017536642029881477, + "learning_rate": 0.0006, + "loss": 4.070173263549805, + "step": 6311 + }, + { + "epoch": 87.6707423580786, + "grad_norm": 0.017303580418229103, + "learning_rate": 0.0006, + "loss": 4.119133949279785, + "step": 6312 + }, + { + "epoch": 87.68471615720524, + "grad_norm": 0.016148580238223076, + "learning_rate": 0.0006, + "loss": 3.994354486465454, + "step": 6313 + }, + { + "epoch": 87.69868995633188, + "grad_norm": 0.01706942543387413, + "learning_rate": 0.0006, + "loss": 4.183113098144531, + "step": 6314 + }, + { + "epoch": 87.71266375545852, + "grad_norm": 0.016971010714769363, + "learning_rate": 0.0006, + "loss": 4.104086875915527, + "step": 6315 + }, + { + "epoch": 87.72663755458515, + "grad_norm": 0.015516847372055054, + "learning_rate": 0.0006, + "loss": 4.2409772872924805, + "step": 6316 + }, + { + "epoch": 87.74061135371178, + "grad_norm": 0.015352269634604454, + "learning_rate": 0.0006, + "loss": 4.026427268981934, + "step": 6317 + }, + { + "epoch": 87.75458515283843, + "grad_norm": 0.017702415585517883, + "learning_rate": 0.0006, + "loss": 4.177840232849121, + "step": 6318 + }, + { + "epoch": 87.76855895196506, + "grad_norm": 0.01629541628062725, + "learning_rate": 0.0006, + "loss": 4.229039192199707, + "step": 6319 + }, + { + "epoch": 87.78253275109171, + "grad_norm": 0.01503711100667715, + "learning_rate": 0.0006, + "loss": 4.214350700378418, + "step": 6320 + }, + { + "epoch": 87.79650655021834, + "grad_norm": 0.014913883060216904, + "learning_rate": 0.0006, + "loss": 4.168880939483643, + "step": 6321 + }, + { + "epoch": 87.81048034934499, + "grad_norm": 0.014513909816741943, + "learning_rate": 0.0006, + "loss": 4.162267684936523, + "step": 6322 + }, + { + "epoch": 87.82445414847162, + "grad_norm": 0.014635873027145863, + "learning_rate": 0.0006, + "loss": 4.220067977905273, + "step": 6323 + }, + { + "epoch": 87.83842794759825, + "grad_norm": 0.017102796584367752, + "learning_rate": 0.0006, + "loss": 4.1493377685546875, + "step": 6324 + }, + { + "epoch": 87.8524017467249, + "grad_norm": 0.015533884055912495, + "learning_rate": 0.0006, + "loss": 4.0646209716796875, + "step": 6325 + }, + { + "epoch": 87.86637554585153, + "grad_norm": 0.014617550186812878, + "learning_rate": 0.0006, + "loss": 4.1390533447265625, + "step": 6326 + }, + { + "epoch": 87.88034934497817, + "grad_norm": 0.014142967760562897, + "learning_rate": 0.0006, + "loss": 4.133934020996094, + "step": 6327 + }, + { + "epoch": 87.8943231441048, + "grad_norm": 0.013182433322072029, + "learning_rate": 0.0006, + "loss": 4.125327110290527, + "step": 6328 + }, + { + "epoch": 87.90829694323143, + "grad_norm": 0.013928229920566082, + "learning_rate": 0.0006, + "loss": 4.142777442932129, + "step": 6329 + }, + { + "epoch": 87.92227074235808, + "grad_norm": 0.015337013639509678, + "learning_rate": 0.0006, + "loss": 4.140210151672363, + "step": 6330 + }, + { + "epoch": 87.93624454148471, + "grad_norm": 0.014873536303639412, + "learning_rate": 0.0006, + "loss": 4.102636337280273, + "step": 6331 + }, + { + "epoch": 87.95021834061136, + "grad_norm": 0.01387431938201189, + "learning_rate": 0.0006, + "loss": 4.149059295654297, + "step": 6332 + }, + { + "epoch": 87.96419213973799, + "grad_norm": 0.013234943151473999, + "learning_rate": 0.0006, + "loss": 4.1564459800720215, + "step": 6333 + }, + { + "epoch": 87.97816593886463, + "grad_norm": 0.013781017623841763, + "learning_rate": 0.0006, + "loss": 4.189839839935303, + "step": 6334 + }, + { + "epoch": 87.99213973799127, + "grad_norm": 0.013733026571571827, + "learning_rate": 0.0006, + "loss": 4.29277229309082, + "step": 6335 + }, + { + "epoch": 88.0, + "grad_norm": 0.014752327464520931, + "learning_rate": 0.0006, + "loss": 4.163867950439453, + "step": 6336 + }, + { + "epoch": 88.0, + "eval_loss": 4.535566329956055, + "eval_runtime": 56.4677, + "eval_samples_per_second": 43.246, + "eval_steps_per_second": 1.364, + "step": 6336 + }, + { + "epoch": 88.01397379912663, + "grad_norm": 0.013958732597529888, + "learning_rate": 0.0006, + "loss": 4.200681209564209, + "step": 6337 + }, + { + "epoch": 88.02794759825328, + "grad_norm": 0.015283545479178429, + "learning_rate": 0.0006, + "loss": 4.1736016273498535, + "step": 6338 + }, + { + "epoch": 88.04192139737991, + "grad_norm": 0.016192059963941574, + "learning_rate": 0.0006, + "loss": 4.040589809417725, + "step": 6339 + }, + { + "epoch": 88.05589519650655, + "grad_norm": 0.018096154555678368, + "learning_rate": 0.0006, + "loss": 4.146696090698242, + "step": 6340 + }, + { + "epoch": 88.06986899563319, + "grad_norm": 0.01797664538025856, + "learning_rate": 0.0006, + "loss": 4.007248878479004, + "step": 6341 + }, + { + "epoch": 88.08384279475983, + "grad_norm": 0.016275612637400627, + "learning_rate": 0.0006, + "loss": 4.193206787109375, + "step": 6342 + }, + { + "epoch": 88.09781659388646, + "grad_norm": 0.014210258610546589, + "learning_rate": 0.0006, + "loss": 4.064548492431641, + "step": 6343 + }, + { + "epoch": 88.1117903930131, + "grad_norm": 0.016085151582956314, + "learning_rate": 0.0006, + "loss": 4.003668785095215, + "step": 6344 + }, + { + "epoch": 88.12576419213974, + "grad_norm": 0.01930483989417553, + "learning_rate": 0.0006, + "loss": 4.236578941345215, + "step": 6345 + }, + { + "epoch": 88.13973799126637, + "grad_norm": 0.017140906304121017, + "learning_rate": 0.0006, + "loss": 3.9593005180358887, + "step": 6346 + }, + { + "epoch": 88.15371179039302, + "grad_norm": 0.01521560549736023, + "learning_rate": 0.0006, + "loss": 4.137644290924072, + "step": 6347 + }, + { + "epoch": 88.16768558951965, + "grad_norm": 0.015788141638040543, + "learning_rate": 0.0006, + "loss": 4.124845027923584, + "step": 6348 + }, + { + "epoch": 88.18165938864628, + "grad_norm": 0.01784597896039486, + "learning_rate": 0.0006, + "loss": 4.139750957489014, + "step": 6349 + }, + { + "epoch": 88.19563318777293, + "grad_norm": 0.016266558319330215, + "learning_rate": 0.0006, + "loss": 4.074851989746094, + "step": 6350 + }, + { + "epoch": 88.20960698689956, + "grad_norm": 0.014024483971297741, + "learning_rate": 0.0006, + "loss": 4.164689064025879, + "step": 6351 + }, + { + "epoch": 88.2235807860262, + "grad_norm": 0.015694420784711838, + "learning_rate": 0.0006, + "loss": 4.219979286193848, + "step": 6352 + }, + { + "epoch": 88.23755458515284, + "grad_norm": 0.014897859655320644, + "learning_rate": 0.0006, + "loss": 4.050738334655762, + "step": 6353 + }, + { + "epoch": 88.25152838427948, + "grad_norm": 0.014827374368906021, + "learning_rate": 0.0006, + "loss": 4.137135028839111, + "step": 6354 + }, + { + "epoch": 88.26550218340611, + "grad_norm": 0.01454879529774189, + "learning_rate": 0.0006, + "loss": 3.9980897903442383, + "step": 6355 + }, + { + "epoch": 88.27947598253274, + "grad_norm": 0.015623947605490685, + "learning_rate": 0.0006, + "loss": 4.226551055908203, + "step": 6356 + }, + { + "epoch": 88.29344978165939, + "grad_norm": 0.015527035109698772, + "learning_rate": 0.0006, + "loss": 4.11053466796875, + "step": 6357 + }, + { + "epoch": 88.30742358078602, + "grad_norm": 0.015242574736475945, + "learning_rate": 0.0006, + "loss": 4.105966567993164, + "step": 6358 + }, + { + "epoch": 88.32139737991267, + "grad_norm": 0.01375538855791092, + "learning_rate": 0.0006, + "loss": 4.148453712463379, + "step": 6359 + }, + { + "epoch": 88.3353711790393, + "grad_norm": 0.01595226116478443, + "learning_rate": 0.0006, + "loss": 4.095836162567139, + "step": 6360 + }, + { + "epoch": 88.34934497816595, + "grad_norm": 0.015252740122377872, + "learning_rate": 0.0006, + "loss": 4.136335849761963, + "step": 6361 + }, + { + "epoch": 88.36331877729258, + "grad_norm": 0.014405866153538227, + "learning_rate": 0.0006, + "loss": 4.084691047668457, + "step": 6362 + }, + { + "epoch": 88.37729257641921, + "grad_norm": 0.014755490235984325, + "learning_rate": 0.0006, + "loss": 4.095475673675537, + "step": 6363 + }, + { + "epoch": 88.39126637554585, + "grad_norm": 0.015339708887040615, + "learning_rate": 0.0006, + "loss": 4.163136959075928, + "step": 6364 + }, + { + "epoch": 88.40524017467249, + "grad_norm": 0.014667164534330368, + "learning_rate": 0.0006, + "loss": 4.062265396118164, + "step": 6365 + }, + { + "epoch": 88.41921397379913, + "grad_norm": 0.014588423073291779, + "learning_rate": 0.0006, + "loss": 4.13359260559082, + "step": 6366 + }, + { + "epoch": 88.43318777292576, + "grad_norm": 0.014005377888679504, + "learning_rate": 0.0006, + "loss": 4.219711780548096, + "step": 6367 + }, + { + "epoch": 88.44716157205241, + "grad_norm": 0.012769930064678192, + "learning_rate": 0.0006, + "loss": 4.114658355712891, + "step": 6368 + }, + { + "epoch": 88.46113537117904, + "grad_norm": 0.013632693327963352, + "learning_rate": 0.0006, + "loss": 4.0675225257873535, + "step": 6369 + }, + { + "epoch": 88.47510917030567, + "grad_norm": 0.013829804956912994, + "learning_rate": 0.0006, + "loss": 4.1125969886779785, + "step": 6370 + }, + { + "epoch": 88.48908296943232, + "grad_norm": 0.013882109895348549, + "learning_rate": 0.0006, + "loss": 4.1627092361450195, + "step": 6371 + }, + { + "epoch": 88.50305676855895, + "grad_norm": 0.014023828320205212, + "learning_rate": 0.0006, + "loss": 4.01038932800293, + "step": 6372 + }, + { + "epoch": 88.5170305676856, + "grad_norm": 0.014840706251561642, + "learning_rate": 0.0006, + "loss": 4.117877960205078, + "step": 6373 + }, + { + "epoch": 88.53100436681223, + "grad_norm": 0.014165930449962616, + "learning_rate": 0.0006, + "loss": 4.1864237785339355, + "step": 6374 + }, + { + "epoch": 88.54497816593886, + "grad_norm": 0.014390201307833195, + "learning_rate": 0.0006, + "loss": 4.1044602394104, + "step": 6375 + }, + { + "epoch": 88.5589519650655, + "grad_norm": 0.015037395060062408, + "learning_rate": 0.0006, + "loss": 4.101107120513916, + "step": 6376 + }, + { + "epoch": 88.57292576419214, + "grad_norm": 0.01671452447772026, + "learning_rate": 0.0006, + "loss": 4.170581817626953, + "step": 6377 + }, + { + "epoch": 88.58689956331878, + "grad_norm": 0.01993076130747795, + "learning_rate": 0.0006, + "loss": 4.261412620544434, + "step": 6378 + }, + { + "epoch": 88.60087336244541, + "grad_norm": 0.023721016943454742, + "learning_rate": 0.0006, + "loss": 4.199363708496094, + "step": 6379 + }, + { + "epoch": 88.61484716157206, + "grad_norm": 0.023726163432002068, + "learning_rate": 0.0006, + "loss": 4.061285018920898, + "step": 6380 + }, + { + "epoch": 88.62882096069869, + "grad_norm": 0.018352877348661423, + "learning_rate": 0.0006, + "loss": 4.0721282958984375, + "step": 6381 + }, + { + "epoch": 88.64279475982532, + "grad_norm": 0.01679830253124237, + "learning_rate": 0.0006, + "loss": 4.1717939376831055, + "step": 6382 + }, + { + "epoch": 88.65676855895197, + "grad_norm": 0.017193637788295746, + "learning_rate": 0.0006, + "loss": 4.230783939361572, + "step": 6383 + }, + { + "epoch": 88.6707423580786, + "grad_norm": 0.01581464521586895, + "learning_rate": 0.0006, + "loss": 4.029906272888184, + "step": 6384 + }, + { + "epoch": 88.68471615720524, + "grad_norm": 0.014049704186618328, + "learning_rate": 0.0006, + "loss": 4.192776679992676, + "step": 6385 + }, + { + "epoch": 88.69868995633188, + "grad_norm": 0.017214450985193253, + "learning_rate": 0.0006, + "loss": 4.248916149139404, + "step": 6386 + }, + { + "epoch": 88.71266375545852, + "grad_norm": 0.0173103716224432, + "learning_rate": 0.0006, + "loss": 4.124347686767578, + "step": 6387 + }, + { + "epoch": 88.72663755458515, + "grad_norm": 0.015745092183351517, + "learning_rate": 0.0006, + "loss": 4.192878246307373, + "step": 6388 + }, + { + "epoch": 88.74061135371178, + "grad_norm": 0.013750772923231125, + "learning_rate": 0.0006, + "loss": 4.1564483642578125, + "step": 6389 + }, + { + "epoch": 88.75458515283843, + "grad_norm": 0.014969858340919018, + "learning_rate": 0.0006, + "loss": 4.0980424880981445, + "step": 6390 + }, + { + "epoch": 88.76855895196506, + "grad_norm": 0.01287173479795456, + "learning_rate": 0.0006, + "loss": 4.145712852478027, + "step": 6391 + }, + { + "epoch": 88.78253275109171, + "grad_norm": 0.012716300785541534, + "learning_rate": 0.0006, + "loss": 4.214226722717285, + "step": 6392 + }, + { + "epoch": 88.79650655021834, + "grad_norm": 0.013855204917490482, + "learning_rate": 0.0006, + "loss": 4.1667256355285645, + "step": 6393 + }, + { + "epoch": 88.81048034934499, + "grad_norm": 0.013650864362716675, + "learning_rate": 0.0006, + "loss": 4.161942481994629, + "step": 6394 + }, + { + "epoch": 88.82445414847162, + "grad_norm": 0.013780126348137856, + "learning_rate": 0.0006, + "loss": 4.087825775146484, + "step": 6395 + }, + { + "epoch": 88.83842794759825, + "grad_norm": 0.014029948972165585, + "learning_rate": 0.0006, + "loss": 4.142813205718994, + "step": 6396 + }, + { + "epoch": 88.8524017467249, + "grad_norm": 0.01322692260146141, + "learning_rate": 0.0006, + "loss": 4.106776714324951, + "step": 6397 + }, + { + "epoch": 88.86637554585153, + "grad_norm": 0.013522760011255741, + "learning_rate": 0.0006, + "loss": 4.090996265411377, + "step": 6398 + }, + { + "epoch": 88.88034934497817, + "grad_norm": 0.014474684372544289, + "learning_rate": 0.0006, + "loss": 4.124624252319336, + "step": 6399 + }, + { + "epoch": 88.8943231441048, + "grad_norm": 0.014490422792732716, + "learning_rate": 0.0006, + "loss": 4.065431594848633, + "step": 6400 + }, + { + "epoch": 88.90829694323143, + "grad_norm": 0.015379710122942924, + "learning_rate": 0.0006, + "loss": 4.207939624786377, + "step": 6401 + }, + { + "epoch": 88.92227074235808, + "grad_norm": 0.015914618968963623, + "learning_rate": 0.0006, + "loss": 4.041152000427246, + "step": 6402 + }, + { + "epoch": 88.93624454148471, + "grad_norm": 0.01480048056691885, + "learning_rate": 0.0006, + "loss": 4.138754844665527, + "step": 6403 + }, + { + "epoch": 88.95021834061136, + "grad_norm": 0.013734851032495499, + "learning_rate": 0.0006, + "loss": 4.25323486328125, + "step": 6404 + }, + { + "epoch": 88.96419213973799, + "grad_norm": 0.01452311035245657, + "learning_rate": 0.0006, + "loss": 4.115228176116943, + "step": 6405 + }, + { + "epoch": 88.97816593886463, + "grad_norm": 0.016197681427001953, + "learning_rate": 0.0006, + "loss": 4.099531173706055, + "step": 6406 + }, + { + "epoch": 88.99213973799127, + "grad_norm": 0.017169523984193802, + "learning_rate": 0.0006, + "loss": 4.182981491088867, + "step": 6407 + }, + { + "epoch": 89.0, + "grad_norm": 0.017855646088719368, + "learning_rate": 0.0006, + "loss": 4.1470136642456055, + "step": 6408 + }, + { + "epoch": 89.0, + "eval_loss": 4.6035614013671875, + "eval_runtime": 57.0625, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6408 + }, + { + "epoch": 89.01397379912663, + "grad_norm": 0.014834101311862469, + "learning_rate": 0.0006, + "loss": 4.159243583679199, + "step": 6409 + }, + { + "epoch": 89.02794759825328, + "grad_norm": 0.015316772274672985, + "learning_rate": 0.0006, + "loss": 4.0709123611450195, + "step": 6410 + }, + { + "epoch": 89.04192139737991, + "grad_norm": 0.01576591096818447, + "learning_rate": 0.0006, + "loss": 4.0921196937561035, + "step": 6411 + }, + { + "epoch": 89.05589519650655, + "grad_norm": 0.015948045998811722, + "learning_rate": 0.0006, + "loss": 4.181877136230469, + "step": 6412 + }, + { + "epoch": 89.06986899563319, + "grad_norm": 0.016612380743026733, + "learning_rate": 0.0006, + "loss": 4.081249237060547, + "step": 6413 + }, + { + "epoch": 89.08384279475983, + "grad_norm": 0.017455685883760452, + "learning_rate": 0.0006, + "loss": 4.175990104675293, + "step": 6414 + }, + { + "epoch": 89.09781659388646, + "grad_norm": 0.015040521509945393, + "learning_rate": 0.0006, + "loss": 4.121378421783447, + "step": 6415 + }, + { + "epoch": 89.1117903930131, + "grad_norm": 0.014513209462165833, + "learning_rate": 0.0006, + "loss": 3.996107816696167, + "step": 6416 + }, + { + "epoch": 89.12576419213974, + "grad_norm": 0.015455638989806175, + "learning_rate": 0.0006, + "loss": 4.073675155639648, + "step": 6417 + }, + { + "epoch": 89.13973799126637, + "grad_norm": 0.015468129888176918, + "learning_rate": 0.0006, + "loss": 4.130242824554443, + "step": 6418 + }, + { + "epoch": 89.15371179039302, + "grad_norm": 0.015284297987818718, + "learning_rate": 0.0006, + "loss": 4.18555212020874, + "step": 6419 + }, + { + "epoch": 89.16768558951965, + "grad_norm": 0.01476874016225338, + "learning_rate": 0.0006, + "loss": 4.134175777435303, + "step": 6420 + }, + { + "epoch": 89.18165938864628, + "grad_norm": 0.015836201608181, + "learning_rate": 0.0006, + "loss": 4.094529151916504, + "step": 6421 + }, + { + "epoch": 89.19563318777293, + "grad_norm": 0.0161928478628397, + "learning_rate": 0.0006, + "loss": 4.12647819519043, + "step": 6422 + }, + { + "epoch": 89.20960698689956, + "grad_norm": 0.015680814161896706, + "learning_rate": 0.0006, + "loss": 3.9620392322540283, + "step": 6423 + }, + { + "epoch": 89.2235807860262, + "grad_norm": 0.01584681309759617, + "learning_rate": 0.0006, + "loss": 4.049304008483887, + "step": 6424 + }, + { + "epoch": 89.23755458515284, + "grad_norm": 0.01452907919883728, + "learning_rate": 0.0006, + "loss": 4.064027786254883, + "step": 6425 + }, + { + "epoch": 89.25152838427948, + "grad_norm": 0.014002552255988121, + "learning_rate": 0.0006, + "loss": 4.151862621307373, + "step": 6426 + }, + { + "epoch": 89.26550218340611, + "grad_norm": 0.014658625237643719, + "learning_rate": 0.0006, + "loss": 4.1672210693359375, + "step": 6427 + }, + { + "epoch": 89.27947598253274, + "grad_norm": 0.015021766535937786, + "learning_rate": 0.0006, + "loss": 3.9742586612701416, + "step": 6428 + }, + { + "epoch": 89.29344978165939, + "grad_norm": 0.014681321568787098, + "learning_rate": 0.0006, + "loss": 4.024159908294678, + "step": 6429 + }, + { + "epoch": 89.30742358078602, + "grad_norm": 0.014161549508571625, + "learning_rate": 0.0006, + "loss": 4.105640888214111, + "step": 6430 + }, + { + "epoch": 89.32139737991267, + "grad_norm": 0.014801453799009323, + "learning_rate": 0.0006, + "loss": 4.123049736022949, + "step": 6431 + }, + { + "epoch": 89.3353711790393, + "grad_norm": 0.014307181350886822, + "learning_rate": 0.0006, + "loss": 4.086954593658447, + "step": 6432 + }, + { + "epoch": 89.34934497816595, + "grad_norm": 0.013299369253218174, + "learning_rate": 0.0006, + "loss": 4.109533786773682, + "step": 6433 + }, + { + "epoch": 89.36331877729258, + "grad_norm": 0.014388669282197952, + "learning_rate": 0.0006, + "loss": 4.222370147705078, + "step": 6434 + }, + { + "epoch": 89.37729257641921, + "grad_norm": 0.014501879923045635, + "learning_rate": 0.0006, + "loss": 4.224178314208984, + "step": 6435 + }, + { + "epoch": 89.39126637554585, + "grad_norm": 0.014109021984040737, + "learning_rate": 0.0006, + "loss": 4.112702369689941, + "step": 6436 + }, + { + "epoch": 89.40524017467249, + "grad_norm": 0.013521002605557442, + "learning_rate": 0.0006, + "loss": 4.010124683380127, + "step": 6437 + }, + { + "epoch": 89.41921397379913, + "grad_norm": 0.013288144953548908, + "learning_rate": 0.0006, + "loss": 4.127592086791992, + "step": 6438 + }, + { + "epoch": 89.43318777292576, + "grad_norm": 0.013882486149668694, + "learning_rate": 0.0006, + "loss": 4.031871318817139, + "step": 6439 + }, + { + "epoch": 89.44716157205241, + "grad_norm": 0.01289335172623396, + "learning_rate": 0.0006, + "loss": 4.0757551193237305, + "step": 6440 + }, + { + "epoch": 89.46113537117904, + "grad_norm": 0.01356302760541439, + "learning_rate": 0.0006, + "loss": 4.075678825378418, + "step": 6441 + }, + { + "epoch": 89.47510917030567, + "grad_norm": 0.01447434164583683, + "learning_rate": 0.0006, + "loss": 4.093453407287598, + "step": 6442 + }, + { + "epoch": 89.48908296943232, + "grad_norm": 0.014443481341004372, + "learning_rate": 0.0006, + "loss": 4.075372219085693, + "step": 6443 + }, + { + "epoch": 89.50305676855895, + "grad_norm": 0.01314218994230032, + "learning_rate": 0.0006, + "loss": 4.171533584594727, + "step": 6444 + }, + { + "epoch": 89.5170305676856, + "grad_norm": 0.01348346658051014, + "learning_rate": 0.0006, + "loss": 4.227236270904541, + "step": 6445 + }, + { + "epoch": 89.53100436681223, + "grad_norm": 0.013370679691433907, + "learning_rate": 0.0006, + "loss": 3.973062515258789, + "step": 6446 + }, + { + "epoch": 89.54497816593886, + "grad_norm": 0.01410642359405756, + "learning_rate": 0.0006, + "loss": 4.231175422668457, + "step": 6447 + }, + { + "epoch": 89.5589519650655, + "grad_norm": 0.01419337373226881, + "learning_rate": 0.0006, + "loss": 4.145868301391602, + "step": 6448 + }, + { + "epoch": 89.57292576419214, + "grad_norm": 0.013691053725779057, + "learning_rate": 0.0006, + "loss": 4.175131797790527, + "step": 6449 + }, + { + "epoch": 89.58689956331878, + "grad_norm": 0.013714492321014404, + "learning_rate": 0.0006, + "loss": 4.251893997192383, + "step": 6450 + }, + { + "epoch": 89.60087336244541, + "grad_norm": 0.01393197476863861, + "learning_rate": 0.0006, + "loss": 4.047079086303711, + "step": 6451 + }, + { + "epoch": 89.61484716157206, + "grad_norm": 0.013957309536635876, + "learning_rate": 0.0006, + "loss": 4.128530502319336, + "step": 6452 + }, + { + "epoch": 89.62882096069869, + "grad_norm": 0.013346714898943901, + "learning_rate": 0.0006, + "loss": 4.2296929359436035, + "step": 6453 + }, + { + "epoch": 89.64279475982532, + "grad_norm": 0.014243829064071178, + "learning_rate": 0.0006, + "loss": 4.128233909606934, + "step": 6454 + }, + { + "epoch": 89.65676855895197, + "grad_norm": 0.01606106385588646, + "learning_rate": 0.0006, + "loss": 4.202202320098877, + "step": 6455 + }, + { + "epoch": 89.6707423580786, + "grad_norm": 0.016074348241090775, + "learning_rate": 0.0006, + "loss": 4.215323448181152, + "step": 6456 + }, + { + "epoch": 89.68471615720524, + "grad_norm": 0.01611054316163063, + "learning_rate": 0.0006, + "loss": 4.096758842468262, + "step": 6457 + }, + { + "epoch": 89.69868995633188, + "grad_norm": 0.013788217678666115, + "learning_rate": 0.0006, + "loss": 4.098560810089111, + "step": 6458 + }, + { + "epoch": 89.71266375545852, + "grad_norm": 0.014408969320356846, + "learning_rate": 0.0006, + "loss": 4.107398509979248, + "step": 6459 + }, + { + "epoch": 89.72663755458515, + "grad_norm": 0.014765267260372639, + "learning_rate": 0.0006, + "loss": 4.057718276977539, + "step": 6460 + }, + { + "epoch": 89.74061135371178, + "grad_norm": 0.017395280301570892, + "learning_rate": 0.0006, + "loss": 4.100795269012451, + "step": 6461 + }, + { + "epoch": 89.75458515283843, + "grad_norm": 0.016821149736642838, + "learning_rate": 0.0006, + "loss": 4.153731346130371, + "step": 6462 + }, + { + "epoch": 89.76855895196506, + "grad_norm": 0.015633124858140945, + "learning_rate": 0.0006, + "loss": 4.061585903167725, + "step": 6463 + }, + { + "epoch": 89.78253275109171, + "grad_norm": 0.01451497245579958, + "learning_rate": 0.0006, + "loss": 4.1624884605407715, + "step": 6464 + }, + { + "epoch": 89.79650655021834, + "grad_norm": 0.013868392445147038, + "learning_rate": 0.0006, + "loss": 4.178536415100098, + "step": 6465 + }, + { + "epoch": 89.81048034934499, + "grad_norm": 0.014259875752031803, + "learning_rate": 0.0006, + "loss": 4.073114395141602, + "step": 6466 + }, + { + "epoch": 89.82445414847162, + "grad_norm": 0.013651594519615173, + "learning_rate": 0.0006, + "loss": 4.050265312194824, + "step": 6467 + }, + { + "epoch": 89.83842794759825, + "grad_norm": 0.014063458889722824, + "learning_rate": 0.0006, + "loss": 4.2014689445495605, + "step": 6468 + }, + { + "epoch": 89.8524017467249, + "grad_norm": 0.014179578050971031, + "learning_rate": 0.0006, + "loss": 4.110550880432129, + "step": 6469 + }, + { + "epoch": 89.86637554585153, + "grad_norm": 0.013616854324936867, + "learning_rate": 0.0006, + "loss": 4.1776580810546875, + "step": 6470 + }, + { + "epoch": 89.88034934497817, + "grad_norm": 0.014351330697536469, + "learning_rate": 0.0006, + "loss": 4.058526992797852, + "step": 6471 + }, + { + "epoch": 89.8943231441048, + "grad_norm": 0.0141244325786829, + "learning_rate": 0.0006, + "loss": 4.068673610687256, + "step": 6472 + }, + { + "epoch": 89.90829694323143, + "grad_norm": 0.01552131213247776, + "learning_rate": 0.0006, + "loss": 4.086787223815918, + "step": 6473 + }, + { + "epoch": 89.92227074235808, + "grad_norm": 0.014614894054830074, + "learning_rate": 0.0006, + "loss": 4.2190656661987305, + "step": 6474 + }, + { + "epoch": 89.93624454148471, + "grad_norm": 0.014369758777320385, + "learning_rate": 0.0006, + "loss": 4.117021560668945, + "step": 6475 + }, + { + "epoch": 89.95021834061136, + "grad_norm": 0.015352046117186546, + "learning_rate": 0.0006, + "loss": 4.167664527893066, + "step": 6476 + }, + { + "epoch": 89.96419213973799, + "grad_norm": 0.014314834959805012, + "learning_rate": 0.0006, + "loss": 4.160182952880859, + "step": 6477 + }, + { + "epoch": 89.97816593886463, + "grad_norm": 0.013511341996490955, + "learning_rate": 0.0006, + "loss": 4.096034049987793, + "step": 6478 + }, + { + "epoch": 89.99213973799127, + "grad_norm": 0.01308482326567173, + "learning_rate": 0.0006, + "loss": 4.062689781188965, + "step": 6479 + }, + { + "epoch": 90.0, + "grad_norm": 0.01442806702107191, + "learning_rate": 0.0006, + "loss": 4.1955885887146, + "step": 6480 + }, + { + "epoch": 90.0, + "eval_loss": 4.644347667694092, + "eval_runtime": 56.8677, + "eval_samples_per_second": 42.942, + "eval_steps_per_second": 1.354, + "step": 6480 + }, + { + "epoch": 90.01397379912663, + "grad_norm": 0.014839007519185543, + "learning_rate": 0.0006, + "loss": 4.171868324279785, + "step": 6481 + }, + { + "epoch": 90.02794759825328, + "grad_norm": 0.01476900465786457, + "learning_rate": 0.0006, + "loss": 4.119754791259766, + "step": 6482 + }, + { + "epoch": 90.04192139737991, + "grad_norm": 0.015179301612079144, + "learning_rate": 0.0006, + "loss": 4.06740140914917, + "step": 6483 + }, + { + "epoch": 90.05589519650655, + "grad_norm": 0.015552980825304985, + "learning_rate": 0.0006, + "loss": 4.112268447875977, + "step": 6484 + }, + { + "epoch": 90.06986899563319, + "grad_norm": 0.014930234290659428, + "learning_rate": 0.0006, + "loss": 4.071866989135742, + "step": 6485 + }, + { + "epoch": 90.08384279475983, + "grad_norm": 0.014124148525297642, + "learning_rate": 0.0006, + "loss": 4.013731956481934, + "step": 6486 + }, + { + "epoch": 90.09781659388646, + "grad_norm": 0.014204252511262894, + "learning_rate": 0.0006, + "loss": 4.133620738983154, + "step": 6487 + }, + { + "epoch": 90.1117903930131, + "grad_norm": 0.014845543541014194, + "learning_rate": 0.0006, + "loss": 4.122956275939941, + "step": 6488 + }, + { + "epoch": 90.12576419213974, + "grad_norm": 0.0146744754165411, + "learning_rate": 0.0006, + "loss": 4.087121963500977, + "step": 6489 + }, + { + "epoch": 90.13973799126637, + "grad_norm": 0.016007190570235252, + "learning_rate": 0.0006, + "loss": 3.98740816116333, + "step": 6490 + }, + { + "epoch": 90.15371179039302, + "grad_norm": 0.018072661012411118, + "learning_rate": 0.0006, + "loss": 4.1155009269714355, + "step": 6491 + }, + { + "epoch": 90.16768558951965, + "grad_norm": 0.016868166625499725, + "learning_rate": 0.0006, + "loss": 4.123727321624756, + "step": 6492 + }, + { + "epoch": 90.18165938864628, + "grad_norm": 0.017091726884245872, + "learning_rate": 0.0006, + "loss": 4.1405229568481445, + "step": 6493 + }, + { + "epoch": 90.19563318777293, + "grad_norm": 0.018329855054616928, + "learning_rate": 0.0006, + "loss": 4.093900680541992, + "step": 6494 + }, + { + "epoch": 90.20960698689956, + "grad_norm": 0.018906457349658012, + "learning_rate": 0.0006, + "loss": 4.0552849769592285, + "step": 6495 + }, + { + "epoch": 90.2235807860262, + "grad_norm": 0.017275869846343994, + "learning_rate": 0.0006, + "loss": 4.1310625076293945, + "step": 6496 + }, + { + "epoch": 90.23755458515284, + "grad_norm": 0.015029897913336754, + "learning_rate": 0.0006, + "loss": 4.100166320800781, + "step": 6497 + }, + { + "epoch": 90.25152838427948, + "grad_norm": 0.014891199767589569, + "learning_rate": 0.0006, + "loss": 4.1314802169799805, + "step": 6498 + }, + { + "epoch": 90.26550218340611, + "grad_norm": 0.014877690002322197, + "learning_rate": 0.0006, + "loss": 4.076098442077637, + "step": 6499 + }, + { + "epoch": 90.27947598253274, + "grad_norm": 0.014670864678919315, + "learning_rate": 0.0006, + "loss": 4.066137313842773, + "step": 6500 + }, + { + "epoch": 90.29344978165939, + "grad_norm": 0.01444234699010849, + "learning_rate": 0.0006, + "loss": 4.013591766357422, + "step": 6501 + }, + { + "epoch": 90.30742358078602, + "grad_norm": 0.01339263841509819, + "learning_rate": 0.0006, + "loss": 4.117959499359131, + "step": 6502 + }, + { + "epoch": 90.32139737991267, + "grad_norm": 0.014753810130059719, + "learning_rate": 0.0006, + "loss": 4.042947769165039, + "step": 6503 + }, + { + "epoch": 90.3353711790393, + "grad_norm": 0.014947964809834957, + "learning_rate": 0.0006, + "loss": 4.112810134887695, + "step": 6504 + }, + { + "epoch": 90.34934497816595, + "grad_norm": 0.015155353583395481, + "learning_rate": 0.0006, + "loss": 4.213214874267578, + "step": 6505 + }, + { + "epoch": 90.36331877729258, + "grad_norm": 0.015502216294407845, + "learning_rate": 0.0006, + "loss": 4.127619743347168, + "step": 6506 + }, + { + "epoch": 90.37729257641921, + "grad_norm": 0.015377781353890896, + "learning_rate": 0.0006, + "loss": 4.135862350463867, + "step": 6507 + }, + { + "epoch": 90.39126637554585, + "grad_norm": 0.01662006787955761, + "learning_rate": 0.0006, + "loss": 4.076877593994141, + "step": 6508 + }, + { + "epoch": 90.40524017467249, + "grad_norm": 0.016324395313858986, + "learning_rate": 0.0006, + "loss": 4.214895725250244, + "step": 6509 + }, + { + "epoch": 90.41921397379913, + "grad_norm": 0.016099024564027786, + "learning_rate": 0.0006, + "loss": 4.139717102050781, + "step": 6510 + }, + { + "epoch": 90.43318777292576, + "grad_norm": 0.014403503388166428, + "learning_rate": 0.0006, + "loss": 4.040343284606934, + "step": 6511 + }, + { + "epoch": 90.44716157205241, + "grad_norm": 0.01413416862487793, + "learning_rate": 0.0006, + "loss": 4.087702751159668, + "step": 6512 + }, + { + "epoch": 90.46113537117904, + "grad_norm": 0.01603989489376545, + "learning_rate": 0.0006, + "loss": 4.173558712005615, + "step": 6513 + }, + { + "epoch": 90.47510917030567, + "grad_norm": 0.016846206039190292, + "learning_rate": 0.0006, + "loss": 4.1955461502075195, + "step": 6514 + }, + { + "epoch": 90.48908296943232, + "grad_norm": 0.01652812585234642, + "learning_rate": 0.0006, + "loss": 4.061838150024414, + "step": 6515 + }, + { + "epoch": 90.50305676855895, + "grad_norm": 0.015100528486073017, + "learning_rate": 0.0006, + "loss": 4.041808605194092, + "step": 6516 + }, + { + "epoch": 90.5170305676856, + "grad_norm": 0.013867822475731373, + "learning_rate": 0.0006, + "loss": 4.12662410736084, + "step": 6517 + }, + { + "epoch": 90.53100436681223, + "grad_norm": 0.013684107922017574, + "learning_rate": 0.0006, + "loss": 4.1254048347473145, + "step": 6518 + }, + { + "epoch": 90.54497816593886, + "grad_norm": 0.014200013130903244, + "learning_rate": 0.0006, + "loss": 4.183216094970703, + "step": 6519 + }, + { + "epoch": 90.5589519650655, + "grad_norm": 0.013707022182643414, + "learning_rate": 0.0006, + "loss": 4.07176399230957, + "step": 6520 + }, + { + "epoch": 90.57292576419214, + "grad_norm": 0.013996582478284836, + "learning_rate": 0.0006, + "loss": 4.006575584411621, + "step": 6521 + }, + { + "epoch": 90.58689956331878, + "grad_norm": 0.014886210672557354, + "learning_rate": 0.0006, + "loss": 4.118956565856934, + "step": 6522 + }, + { + "epoch": 90.60087336244541, + "grad_norm": 0.014022685587406158, + "learning_rate": 0.0006, + "loss": 3.96986722946167, + "step": 6523 + }, + { + "epoch": 90.61484716157206, + "grad_norm": 0.014547634869813919, + "learning_rate": 0.0006, + "loss": 4.122886657714844, + "step": 6524 + }, + { + "epoch": 90.62882096069869, + "grad_norm": 0.013682851567864418, + "learning_rate": 0.0006, + "loss": 3.988292694091797, + "step": 6525 + }, + { + "epoch": 90.64279475982532, + "grad_norm": 0.014016421511769295, + "learning_rate": 0.0006, + "loss": 4.110270977020264, + "step": 6526 + }, + { + "epoch": 90.65676855895197, + "grad_norm": 0.013392233289778233, + "learning_rate": 0.0006, + "loss": 4.065059661865234, + "step": 6527 + }, + { + "epoch": 90.6707423580786, + "grad_norm": 0.012107795104384422, + "learning_rate": 0.0006, + "loss": 4.034260272979736, + "step": 6528 + }, + { + "epoch": 90.68471615720524, + "grad_norm": 0.013198604807257652, + "learning_rate": 0.0006, + "loss": 4.162083625793457, + "step": 6529 + }, + { + "epoch": 90.69868995633188, + "grad_norm": 0.012380579486489296, + "learning_rate": 0.0006, + "loss": 4.040456295013428, + "step": 6530 + }, + { + "epoch": 90.71266375545852, + "grad_norm": 0.012784135527908802, + "learning_rate": 0.0006, + "loss": 4.0823798179626465, + "step": 6531 + }, + { + "epoch": 90.72663755458515, + "grad_norm": 0.012920220382511616, + "learning_rate": 0.0006, + "loss": 4.181064128875732, + "step": 6532 + }, + { + "epoch": 90.74061135371178, + "grad_norm": 0.01338814664632082, + "learning_rate": 0.0006, + "loss": 4.111660003662109, + "step": 6533 + }, + { + "epoch": 90.75458515283843, + "grad_norm": 0.013628040440380573, + "learning_rate": 0.0006, + "loss": 4.237852573394775, + "step": 6534 + }, + { + "epoch": 90.76855895196506, + "grad_norm": 0.013384854421019554, + "learning_rate": 0.0006, + "loss": 4.121862888336182, + "step": 6535 + }, + { + "epoch": 90.78253275109171, + "grad_norm": 0.014726929366588593, + "learning_rate": 0.0006, + "loss": 4.1695990562438965, + "step": 6536 + }, + { + "epoch": 90.79650655021834, + "grad_norm": 0.0161006897687912, + "learning_rate": 0.0006, + "loss": 4.17794132232666, + "step": 6537 + }, + { + "epoch": 90.81048034934499, + "grad_norm": 0.016761884093284607, + "learning_rate": 0.0006, + "loss": 4.093764781951904, + "step": 6538 + }, + { + "epoch": 90.82445414847162, + "grad_norm": 0.017246825620532036, + "learning_rate": 0.0006, + "loss": 3.9382331371307373, + "step": 6539 + }, + { + "epoch": 90.83842794759825, + "grad_norm": 0.018141290172934532, + "learning_rate": 0.0006, + "loss": 3.973609209060669, + "step": 6540 + }, + { + "epoch": 90.8524017467249, + "grad_norm": 0.018129250034689903, + "learning_rate": 0.0006, + "loss": 4.227414131164551, + "step": 6541 + }, + { + "epoch": 90.86637554585153, + "grad_norm": 0.018234573304653168, + "learning_rate": 0.0006, + "loss": 4.151870250701904, + "step": 6542 + }, + { + "epoch": 90.88034934497817, + "grad_norm": 0.01560172438621521, + "learning_rate": 0.0006, + "loss": 4.127397537231445, + "step": 6543 + }, + { + "epoch": 90.8943231441048, + "grad_norm": 0.014367681927978992, + "learning_rate": 0.0006, + "loss": 4.17820930480957, + "step": 6544 + }, + { + "epoch": 90.90829694323143, + "grad_norm": 0.016071319580078125, + "learning_rate": 0.0006, + "loss": 4.059854507446289, + "step": 6545 + }, + { + "epoch": 90.92227074235808, + "grad_norm": 0.01589169353246689, + "learning_rate": 0.0006, + "loss": 4.282748222351074, + "step": 6546 + }, + { + "epoch": 90.93624454148471, + "grad_norm": 0.013639772310853004, + "learning_rate": 0.0006, + "loss": 4.147368907928467, + "step": 6547 + }, + { + "epoch": 90.95021834061136, + "grad_norm": 0.014248427003622055, + "learning_rate": 0.0006, + "loss": 4.240248680114746, + "step": 6548 + }, + { + "epoch": 90.96419213973799, + "grad_norm": 0.014600320719182491, + "learning_rate": 0.0006, + "loss": 4.122123718261719, + "step": 6549 + }, + { + "epoch": 90.97816593886463, + "grad_norm": 0.014371675439178944, + "learning_rate": 0.0006, + "loss": 4.013302803039551, + "step": 6550 + }, + { + "epoch": 90.99213973799127, + "grad_norm": 0.013133526779711246, + "learning_rate": 0.0006, + "loss": 4.097136497497559, + "step": 6551 + }, + { + "epoch": 91.0, + "grad_norm": 0.015599353238940239, + "learning_rate": 0.0006, + "loss": 4.198554039001465, + "step": 6552 + }, + { + "epoch": 91.0, + "eval_loss": 4.603907585144043, + "eval_runtime": 56.9418, + "eval_samples_per_second": 42.886, + "eval_steps_per_second": 1.352, + "step": 6552 + }, + { + "epoch": 91.01397379912663, + "grad_norm": 0.015917908400297165, + "learning_rate": 0.0006, + "loss": 4.133877754211426, + "step": 6553 + }, + { + "epoch": 91.02794759825328, + "grad_norm": 0.015485800802707672, + "learning_rate": 0.0006, + "loss": 4.111617088317871, + "step": 6554 + }, + { + "epoch": 91.04192139737991, + "grad_norm": 0.012969603762030602, + "learning_rate": 0.0006, + "loss": 4.064828395843506, + "step": 6555 + }, + { + "epoch": 91.05589519650655, + "grad_norm": 0.014482097700238228, + "learning_rate": 0.0006, + "loss": 4.101710319519043, + "step": 6556 + }, + { + "epoch": 91.06986899563319, + "grad_norm": 0.017242740839719772, + "learning_rate": 0.0006, + "loss": 4.015342712402344, + "step": 6557 + }, + { + "epoch": 91.08384279475983, + "grad_norm": 0.017100440338253975, + "learning_rate": 0.0006, + "loss": 3.9801602363586426, + "step": 6558 + }, + { + "epoch": 91.09781659388646, + "grad_norm": 0.015969131141901016, + "learning_rate": 0.0006, + "loss": 4.049958229064941, + "step": 6559 + }, + { + "epoch": 91.1117903930131, + "grad_norm": 0.016018927097320557, + "learning_rate": 0.0006, + "loss": 4.094139099121094, + "step": 6560 + }, + { + "epoch": 91.12576419213974, + "grad_norm": 0.018131926655769348, + "learning_rate": 0.0006, + "loss": 3.9499974250793457, + "step": 6561 + }, + { + "epoch": 91.13973799126637, + "grad_norm": 0.019425220787525177, + "learning_rate": 0.0006, + "loss": 3.9910552501678467, + "step": 6562 + }, + { + "epoch": 91.15371179039302, + "grad_norm": 0.01908663846552372, + "learning_rate": 0.0006, + "loss": 4.032179832458496, + "step": 6563 + }, + { + "epoch": 91.16768558951965, + "grad_norm": 0.01871207356452942, + "learning_rate": 0.0006, + "loss": 4.112380027770996, + "step": 6564 + }, + { + "epoch": 91.18165938864628, + "grad_norm": 0.020867953076958656, + "learning_rate": 0.0006, + "loss": 4.06105899810791, + "step": 6565 + }, + { + "epoch": 91.19563318777293, + "grad_norm": 0.019053412601351738, + "learning_rate": 0.0006, + "loss": 3.992504596710205, + "step": 6566 + }, + { + "epoch": 91.20960698689956, + "grad_norm": 0.016084402799606323, + "learning_rate": 0.0006, + "loss": 4.100559234619141, + "step": 6567 + }, + { + "epoch": 91.2235807860262, + "grad_norm": 0.016857590526342392, + "learning_rate": 0.0006, + "loss": 4.1567277908325195, + "step": 6568 + }, + { + "epoch": 91.23755458515284, + "grad_norm": 0.01702957972884178, + "learning_rate": 0.0006, + "loss": 4.234771728515625, + "step": 6569 + }, + { + "epoch": 91.25152838427948, + "grad_norm": 0.016869494691491127, + "learning_rate": 0.0006, + "loss": 3.9931387901306152, + "step": 6570 + }, + { + "epoch": 91.26550218340611, + "grad_norm": 0.01696581020951271, + "learning_rate": 0.0006, + "loss": 4.184908866882324, + "step": 6571 + }, + { + "epoch": 91.27947598253274, + "grad_norm": 0.014982126653194427, + "learning_rate": 0.0006, + "loss": 4.036668300628662, + "step": 6572 + }, + { + "epoch": 91.29344978165939, + "grad_norm": 0.014363301917910576, + "learning_rate": 0.0006, + "loss": 4.074010848999023, + "step": 6573 + }, + { + "epoch": 91.30742358078602, + "grad_norm": 0.015492337755858898, + "learning_rate": 0.0006, + "loss": 4.127626419067383, + "step": 6574 + }, + { + "epoch": 91.32139737991267, + "grad_norm": 0.015375050716102123, + "learning_rate": 0.0006, + "loss": 4.159229755401611, + "step": 6575 + }, + { + "epoch": 91.3353711790393, + "grad_norm": 0.015068565495312214, + "learning_rate": 0.0006, + "loss": 4.172623634338379, + "step": 6576 + }, + { + "epoch": 91.34934497816595, + "grad_norm": 0.0140256742015481, + "learning_rate": 0.0006, + "loss": 4.207065105438232, + "step": 6577 + }, + { + "epoch": 91.36331877729258, + "grad_norm": 0.014220152981579304, + "learning_rate": 0.0006, + "loss": 4.131444931030273, + "step": 6578 + }, + { + "epoch": 91.37729257641921, + "grad_norm": 0.0143149234354496, + "learning_rate": 0.0006, + "loss": 4.065218448638916, + "step": 6579 + }, + { + "epoch": 91.39126637554585, + "grad_norm": 0.013149659149348736, + "learning_rate": 0.0006, + "loss": 4.106743335723877, + "step": 6580 + }, + { + "epoch": 91.40524017467249, + "grad_norm": 0.014550330117344856, + "learning_rate": 0.0006, + "loss": 4.123722553253174, + "step": 6581 + }, + { + "epoch": 91.41921397379913, + "grad_norm": 0.014611710794270039, + "learning_rate": 0.0006, + "loss": 4.079610824584961, + "step": 6582 + }, + { + "epoch": 91.43318777292576, + "grad_norm": 0.013369769789278507, + "learning_rate": 0.0006, + "loss": 4.133187770843506, + "step": 6583 + }, + { + "epoch": 91.44716157205241, + "grad_norm": 0.013243545778095722, + "learning_rate": 0.0006, + "loss": 4.042607307434082, + "step": 6584 + }, + { + "epoch": 91.46113537117904, + "grad_norm": 0.013388266786932945, + "learning_rate": 0.0006, + "loss": 4.14863395690918, + "step": 6585 + }, + { + "epoch": 91.47510917030567, + "grad_norm": 0.014333273284137249, + "learning_rate": 0.0006, + "loss": 4.109699726104736, + "step": 6586 + }, + { + "epoch": 91.48908296943232, + "grad_norm": 0.012742757797241211, + "learning_rate": 0.0006, + "loss": 4.126582145690918, + "step": 6587 + }, + { + "epoch": 91.50305676855895, + "grad_norm": 0.012880304828286171, + "learning_rate": 0.0006, + "loss": 4.205300331115723, + "step": 6588 + }, + { + "epoch": 91.5170305676856, + "grad_norm": 0.013896043412387371, + "learning_rate": 0.0006, + "loss": 4.132510185241699, + "step": 6589 + }, + { + "epoch": 91.53100436681223, + "grad_norm": 0.014361281879246235, + "learning_rate": 0.0006, + "loss": 4.1487932205200195, + "step": 6590 + }, + { + "epoch": 91.54497816593886, + "grad_norm": 0.015137741342186928, + "learning_rate": 0.0006, + "loss": 4.1564531326293945, + "step": 6591 + }, + { + "epoch": 91.5589519650655, + "grad_norm": 0.014660682529211044, + "learning_rate": 0.0006, + "loss": 4.056064128875732, + "step": 6592 + }, + { + "epoch": 91.57292576419214, + "grad_norm": 0.014888476580381393, + "learning_rate": 0.0006, + "loss": 4.0633344650268555, + "step": 6593 + }, + { + "epoch": 91.58689956331878, + "grad_norm": 0.012162433005869389, + "learning_rate": 0.0006, + "loss": 4.026293754577637, + "step": 6594 + }, + { + "epoch": 91.60087336244541, + "grad_norm": 0.012892271392047405, + "learning_rate": 0.0006, + "loss": 4.267086029052734, + "step": 6595 + }, + { + "epoch": 91.61484716157206, + "grad_norm": 0.01450759544968605, + "learning_rate": 0.0006, + "loss": 4.149359703063965, + "step": 6596 + }, + { + "epoch": 91.62882096069869, + "grad_norm": 0.01419843453913927, + "learning_rate": 0.0006, + "loss": 4.094183921813965, + "step": 6597 + }, + { + "epoch": 91.64279475982532, + "grad_norm": 0.014401616528630257, + "learning_rate": 0.0006, + "loss": 4.125172138214111, + "step": 6598 + }, + { + "epoch": 91.65676855895197, + "grad_norm": 0.013945513404905796, + "learning_rate": 0.0006, + "loss": 4.146623134613037, + "step": 6599 + }, + { + "epoch": 91.6707423580786, + "grad_norm": 0.013516047969460487, + "learning_rate": 0.0006, + "loss": 4.160526275634766, + "step": 6600 + }, + { + "epoch": 91.68471615720524, + "grad_norm": 0.013262378051877022, + "learning_rate": 0.0006, + "loss": 4.074923038482666, + "step": 6601 + }, + { + "epoch": 91.69868995633188, + "grad_norm": 0.014365669339895248, + "learning_rate": 0.0006, + "loss": 4.118094444274902, + "step": 6602 + }, + { + "epoch": 91.71266375545852, + "grad_norm": 0.016713734716176987, + "learning_rate": 0.0006, + "loss": 4.021799087524414, + "step": 6603 + }, + { + "epoch": 91.72663755458515, + "grad_norm": 0.01988680474460125, + "learning_rate": 0.0006, + "loss": 4.238436222076416, + "step": 6604 + }, + { + "epoch": 91.74061135371178, + "grad_norm": 0.022749733179807663, + "learning_rate": 0.0006, + "loss": 4.010470390319824, + "step": 6605 + }, + { + "epoch": 91.75458515283843, + "grad_norm": 0.022533461451530457, + "learning_rate": 0.0006, + "loss": 4.152201175689697, + "step": 6606 + }, + { + "epoch": 91.76855895196506, + "grad_norm": 0.02021082676947117, + "learning_rate": 0.0006, + "loss": 4.120572090148926, + "step": 6607 + }, + { + "epoch": 91.78253275109171, + "grad_norm": 0.016298197209835052, + "learning_rate": 0.0006, + "loss": 4.115815162658691, + "step": 6608 + }, + { + "epoch": 91.79650655021834, + "grad_norm": 0.017562786117196083, + "learning_rate": 0.0006, + "loss": 4.109724044799805, + "step": 6609 + }, + { + "epoch": 91.81048034934499, + "grad_norm": 0.018153153359889984, + "learning_rate": 0.0006, + "loss": 4.143759727478027, + "step": 6610 + }, + { + "epoch": 91.82445414847162, + "grad_norm": 0.014762306585907936, + "learning_rate": 0.0006, + "loss": 4.25596809387207, + "step": 6611 + }, + { + "epoch": 91.83842794759825, + "grad_norm": 0.015074343420565128, + "learning_rate": 0.0006, + "loss": 4.207723140716553, + "step": 6612 + }, + { + "epoch": 91.8524017467249, + "grad_norm": 0.015028579160571098, + "learning_rate": 0.0006, + "loss": 4.01950216293335, + "step": 6613 + }, + { + "epoch": 91.86637554585153, + "grad_norm": 0.013996795751154423, + "learning_rate": 0.0006, + "loss": 4.132997989654541, + "step": 6614 + }, + { + "epoch": 91.88034934497817, + "grad_norm": 0.014978236518800259, + "learning_rate": 0.0006, + "loss": 4.132402420043945, + "step": 6615 + }, + { + "epoch": 91.8943231441048, + "grad_norm": 0.013111517764627934, + "learning_rate": 0.0006, + "loss": 4.1223649978637695, + "step": 6616 + }, + { + "epoch": 91.90829694323143, + "grad_norm": 0.013872234150767326, + "learning_rate": 0.0006, + "loss": 4.096590042114258, + "step": 6617 + }, + { + "epoch": 91.92227074235808, + "grad_norm": 0.013642470352351665, + "learning_rate": 0.0006, + "loss": 4.236227035522461, + "step": 6618 + }, + { + "epoch": 91.93624454148471, + "grad_norm": 0.014684873633086681, + "learning_rate": 0.0006, + "loss": 4.037961959838867, + "step": 6619 + }, + { + "epoch": 91.95021834061136, + "grad_norm": 0.014498206786811352, + "learning_rate": 0.0006, + "loss": 4.043325424194336, + "step": 6620 + }, + { + "epoch": 91.96419213973799, + "grad_norm": 0.012768656946718693, + "learning_rate": 0.0006, + "loss": 4.138803482055664, + "step": 6621 + }, + { + "epoch": 91.97816593886463, + "grad_norm": 0.012155905365943909, + "learning_rate": 0.0006, + "loss": 4.163424968719482, + "step": 6622 + }, + { + "epoch": 91.99213973799127, + "grad_norm": 0.012584930285811424, + "learning_rate": 0.0006, + "loss": 4.086629867553711, + "step": 6623 + }, + { + "epoch": 92.0, + "grad_norm": 0.01578151248395443, + "learning_rate": 0.0006, + "loss": 4.108386993408203, + "step": 6624 + }, + { + "epoch": 92.0, + "eval_loss": 4.541166305541992, + "eval_runtime": 56.871, + "eval_samples_per_second": 42.939, + "eval_steps_per_second": 1.354, + "step": 6624 + }, + { + "epoch": 92.01397379912663, + "grad_norm": 0.016901545226573944, + "learning_rate": 0.0006, + "loss": 4.043694496154785, + "step": 6625 + }, + { + "epoch": 92.02794759825328, + "grad_norm": 0.016299039125442505, + "learning_rate": 0.0006, + "loss": 4.100504398345947, + "step": 6626 + }, + { + "epoch": 92.04192139737991, + "grad_norm": 0.013988777063786983, + "learning_rate": 0.0006, + "loss": 4.063943862915039, + "step": 6627 + }, + { + "epoch": 92.05589519650655, + "grad_norm": 0.01343096699565649, + "learning_rate": 0.0006, + "loss": 4.0321197509765625, + "step": 6628 + }, + { + "epoch": 92.06986899563319, + "grad_norm": 0.014343147166073322, + "learning_rate": 0.0006, + "loss": 4.095499038696289, + "step": 6629 + }, + { + "epoch": 92.08384279475983, + "grad_norm": 0.014989756047725677, + "learning_rate": 0.0006, + "loss": 4.122310638427734, + "step": 6630 + }, + { + "epoch": 92.09781659388646, + "grad_norm": 0.013929174281656742, + "learning_rate": 0.0006, + "loss": 4.02870512008667, + "step": 6631 + }, + { + "epoch": 92.1117903930131, + "grad_norm": 0.014139998704195023, + "learning_rate": 0.0006, + "loss": 4.180675506591797, + "step": 6632 + }, + { + "epoch": 92.12576419213974, + "grad_norm": 0.012920474633574486, + "learning_rate": 0.0006, + "loss": 4.111686706542969, + "step": 6633 + }, + { + "epoch": 92.13973799126637, + "grad_norm": 0.01291937381029129, + "learning_rate": 0.0006, + "loss": 3.958092212677002, + "step": 6634 + }, + { + "epoch": 92.15371179039302, + "grad_norm": 0.013250474818050861, + "learning_rate": 0.0006, + "loss": 4.075105667114258, + "step": 6635 + }, + { + "epoch": 92.16768558951965, + "grad_norm": 0.01321406476199627, + "learning_rate": 0.0006, + "loss": 4.072322368621826, + "step": 6636 + }, + { + "epoch": 92.18165938864628, + "grad_norm": 0.014002865180373192, + "learning_rate": 0.0006, + "loss": 4.061572074890137, + "step": 6637 + }, + { + "epoch": 92.19563318777293, + "grad_norm": 0.013876105658710003, + "learning_rate": 0.0006, + "loss": 4.099213600158691, + "step": 6638 + }, + { + "epoch": 92.20960698689956, + "grad_norm": 0.01464459951967001, + "learning_rate": 0.0006, + "loss": 4.1573638916015625, + "step": 6639 + }, + { + "epoch": 92.2235807860262, + "grad_norm": 0.015354706905782223, + "learning_rate": 0.0006, + "loss": 4.078714370727539, + "step": 6640 + }, + { + "epoch": 92.23755458515284, + "grad_norm": 0.015265689231455326, + "learning_rate": 0.0006, + "loss": 4.036652565002441, + "step": 6641 + }, + { + "epoch": 92.25152838427948, + "grad_norm": 0.016140855848789215, + "learning_rate": 0.0006, + "loss": 4.120541095733643, + "step": 6642 + }, + { + "epoch": 92.26550218340611, + "grad_norm": 0.017044052481651306, + "learning_rate": 0.0006, + "loss": 4.015849590301514, + "step": 6643 + }, + { + "epoch": 92.27947598253274, + "grad_norm": 0.01469327136874199, + "learning_rate": 0.0006, + "loss": 4.137286186218262, + "step": 6644 + }, + { + "epoch": 92.29344978165939, + "grad_norm": 0.014596942812204361, + "learning_rate": 0.0006, + "loss": 4.11309814453125, + "step": 6645 + }, + { + "epoch": 92.30742358078602, + "grad_norm": 0.014671559445559978, + "learning_rate": 0.0006, + "loss": 4.087950229644775, + "step": 6646 + }, + { + "epoch": 92.32139737991267, + "grad_norm": 0.015040507540106773, + "learning_rate": 0.0006, + "loss": 3.967053174972534, + "step": 6647 + }, + { + "epoch": 92.3353711790393, + "grad_norm": 0.01417640782892704, + "learning_rate": 0.0006, + "loss": 3.990346908569336, + "step": 6648 + }, + { + "epoch": 92.34934497816595, + "grad_norm": 0.013855187222361565, + "learning_rate": 0.0006, + "loss": 3.965132474899292, + "step": 6649 + }, + { + "epoch": 92.36331877729258, + "grad_norm": 0.014291216619312763, + "learning_rate": 0.0006, + "loss": 4.161717891693115, + "step": 6650 + }, + { + "epoch": 92.37729257641921, + "grad_norm": 0.014148853719234467, + "learning_rate": 0.0006, + "loss": 4.019951820373535, + "step": 6651 + }, + { + "epoch": 92.39126637554585, + "grad_norm": 0.01414303295314312, + "learning_rate": 0.0006, + "loss": 4.125976085662842, + "step": 6652 + }, + { + "epoch": 92.40524017467249, + "grad_norm": 0.01358176488429308, + "learning_rate": 0.0006, + "loss": 4.088022232055664, + "step": 6653 + }, + { + "epoch": 92.41921397379913, + "grad_norm": 0.014174354262650013, + "learning_rate": 0.0006, + "loss": 4.1137518882751465, + "step": 6654 + }, + { + "epoch": 92.43318777292576, + "grad_norm": 0.013756301254034042, + "learning_rate": 0.0006, + "loss": 4.0088725090026855, + "step": 6655 + }, + { + "epoch": 92.44716157205241, + "grad_norm": 0.014257770963013172, + "learning_rate": 0.0006, + "loss": 4.05248498916626, + "step": 6656 + }, + { + "epoch": 92.46113537117904, + "grad_norm": 0.015319934114813805, + "learning_rate": 0.0006, + "loss": 4.1375732421875, + "step": 6657 + }, + { + "epoch": 92.47510917030567, + "grad_norm": 0.014655457809567451, + "learning_rate": 0.0006, + "loss": 4.172857284545898, + "step": 6658 + }, + { + "epoch": 92.48908296943232, + "grad_norm": 0.014981928281486034, + "learning_rate": 0.0006, + "loss": 4.064846038818359, + "step": 6659 + }, + { + "epoch": 92.50305676855895, + "grad_norm": 0.015197387896478176, + "learning_rate": 0.0006, + "loss": 4.119397163391113, + "step": 6660 + }, + { + "epoch": 92.5170305676856, + "grad_norm": 0.016060935333371162, + "learning_rate": 0.0006, + "loss": 4.09212064743042, + "step": 6661 + }, + { + "epoch": 92.53100436681223, + "grad_norm": 0.014884199015796185, + "learning_rate": 0.0006, + "loss": 4.027483940124512, + "step": 6662 + }, + { + "epoch": 92.54497816593886, + "grad_norm": 0.01562194712460041, + "learning_rate": 0.0006, + "loss": 3.9722909927368164, + "step": 6663 + }, + { + "epoch": 92.5589519650655, + "grad_norm": 0.014801043085753918, + "learning_rate": 0.0006, + "loss": 4.048130989074707, + "step": 6664 + }, + { + "epoch": 92.57292576419214, + "grad_norm": 0.013954494148492813, + "learning_rate": 0.0006, + "loss": 4.050351619720459, + "step": 6665 + }, + { + "epoch": 92.58689956331878, + "grad_norm": 0.013292429968714714, + "learning_rate": 0.0006, + "loss": 4.245860576629639, + "step": 6666 + }, + { + "epoch": 92.60087336244541, + "grad_norm": 0.015737095847725868, + "learning_rate": 0.0006, + "loss": 4.154323101043701, + "step": 6667 + }, + { + "epoch": 92.61484716157206, + "grad_norm": 0.018242565914988518, + "learning_rate": 0.0006, + "loss": 4.118710517883301, + "step": 6668 + }, + { + "epoch": 92.62882096069869, + "grad_norm": 0.017666636034846306, + "learning_rate": 0.0006, + "loss": 4.1013875007629395, + "step": 6669 + }, + { + "epoch": 92.64279475982532, + "grad_norm": 0.015017223544418812, + "learning_rate": 0.0006, + "loss": 4.084001064300537, + "step": 6670 + }, + { + "epoch": 92.65676855895197, + "grad_norm": 0.01359238475561142, + "learning_rate": 0.0006, + "loss": 4.091456413269043, + "step": 6671 + }, + { + "epoch": 92.6707423580786, + "grad_norm": 0.014524881727993488, + "learning_rate": 0.0006, + "loss": 4.121245384216309, + "step": 6672 + }, + { + "epoch": 92.68471615720524, + "grad_norm": 0.016214480623602867, + "learning_rate": 0.0006, + "loss": 4.133184432983398, + "step": 6673 + }, + { + "epoch": 92.69868995633188, + "grad_norm": 0.0176643505692482, + "learning_rate": 0.0006, + "loss": 4.142827987670898, + "step": 6674 + }, + { + "epoch": 92.71266375545852, + "grad_norm": 0.017004093155264854, + "learning_rate": 0.0006, + "loss": 4.067666530609131, + "step": 6675 + }, + { + "epoch": 92.72663755458515, + "grad_norm": 0.014077397994697094, + "learning_rate": 0.0006, + "loss": 4.076465129852295, + "step": 6676 + }, + { + "epoch": 92.74061135371178, + "grad_norm": 0.015905627980828285, + "learning_rate": 0.0006, + "loss": 4.083726406097412, + "step": 6677 + }, + { + "epoch": 92.75458515283843, + "grad_norm": 0.016467567533254623, + "learning_rate": 0.0006, + "loss": 4.241217136383057, + "step": 6678 + }, + { + "epoch": 92.76855895196506, + "grad_norm": 0.01541806198656559, + "learning_rate": 0.0006, + "loss": 4.05365514755249, + "step": 6679 + }, + { + "epoch": 92.78253275109171, + "grad_norm": 0.015028242021799088, + "learning_rate": 0.0006, + "loss": 4.029950141906738, + "step": 6680 + }, + { + "epoch": 92.79650655021834, + "grad_norm": 0.014366233721375465, + "learning_rate": 0.0006, + "loss": 4.181772232055664, + "step": 6681 + }, + { + "epoch": 92.81048034934499, + "grad_norm": 0.014313179068267345, + "learning_rate": 0.0006, + "loss": 4.152658462524414, + "step": 6682 + }, + { + "epoch": 92.82445414847162, + "grad_norm": 0.016504302620887756, + "learning_rate": 0.0006, + "loss": 4.080532073974609, + "step": 6683 + }, + { + "epoch": 92.83842794759825, + "grad_norm": 0.01685677096247673, + "learning_rate": 0.0006, + "loss": 4.194268226623535, + "step": 6684 + }, + { + "epoch": 92.8524017467249, + "grad_norm": 0.018684150651097298, + "learning_rate": 0.0006, + "loss": 4.218497276306152, + "step": 6685 + }, + { + "epoch": 92.86637554585153, + "grad_norm": 0.01880732551217079, + "learning_rate": 0.0006, + "loss": 4.133942604064941, + "step": 6686 + }, + { + "epoch": 92.88034934497817, + "grad_norm": 0.016504917293787003, + "learning_rate": 0.0006, + "loss": 4.139945030212402, + "step": 6687 + }, + { + "epoch": 92.8943231441048, + "grad_norm": 0.016495356336236, + "learning_rate": 0.0006, + "loss": 4.098918914794922, + "step": 6688 + }, + { + "epoch": 92.90829694323143, + "grad_norm": 0.0170587245374918, + "learning_rate": 0.0006, + "loss": 4.025818824768066, + "step": 6689 + }, + { + "epoch": 92.92227074235808, + "grad_norm": 0.01718820631504059, + "learning_rate": 0.0006, + "loss": 4.125217437744141, + "step": 6690 + }, + { + "epoch": 92.93624454148471, + "grad_norm": 0.015789354220032692, + "learning_rate": 0.0006, + "loss": 4.150068283081055, + "step": 6691 + }, + { + "epoch": 92.95021834061136, + "grad_norm": 0.01587480492889881, + "learning_rate": 0.0006, + "loss": 4.194794654846191, + "step": 6692 + }, + { + "epoch": 92.96419213973799, + "grad_norm": 0.015669209882616997, + "learning_rate": 0.0006, + "loss": 4.0550336837768555, + "step": 6693 + }, + { + "epoch": 92.97816593886463, + "grad_norm": 0.01519355084747076, + "learning_rate": 0.0006, + "loss": 4.114232063293457, + "step": 6694 + }, + { + "epoch": 92.99213973799127, + "grad_norm": 0.014596261084079742, + "learning_rate": 0.0006, + "loss": 3.992065906524658, + "step": 6695 + }, + { + "epoch": 93.0, + "grad_norm": 0.016378438100218773, + "learning_rate": 0.0006, + "loss": 4.093693733215332, + "step": 6696 + }, + { + "epoch": 93.0, + "eval_loss": 4.541905879974365, + "eval_runtime": 57.0177, + "eval_samples_per_second": 42.829, + "eval_steps_per_second": 1.35, + "step": 6696 + }, + { + "epoch": 93.01397379912663, + "grad_norm": 0.01593143492937088, + "learning_rate": 0.0006, + "loss": 4.052624225616455, + "step": 6697 + }, + { + "epoch": 93.02794759825328, + "grad_norm": 0.016086837276816368, + "learning_rate": 0.0006, + "loss": 4.130306243896484, + "step": 6698 + }, + { + "epoch": 93.04192139737991, + "grad_norm": 0.016004323959350586, + "learning_rate": 0.0006, + "loss": 3.965893268585205, + "step": 6699 + }, + { + "epoch": 93.05589519650655, + "grad_norm": 0.017667284235358238, + "learning_rate": 0.0006, + "loss": 4.095634460449219, + "step": 6700 + }, + { + "epoch": 93.06986899563319, + "grad_norm": 0.01949942484498024, + "learning_rate": 0.0006, + "loss": 3.966041326522827, + "step": 6701 + }, + { + "epoch": 93.08384279475983, + "grad_norm": 0.021898971870541573, + "learning_rate": 0.0006, + "loss": 4.010378837585449, + "step": 6702 + }, + { + "epoch": 93.09781659388646, + "grad_norm": 0.019985264167189598, + "learning_rate": 0.0006, + "loss": 4.061577320098877, + "step": 6703 + }, + { + "epoch": 93.1117903930131, + "grad_norm": 0.016470855101943016, + "learning_rate": 0.0006, + "loss": 4.0601677894592285, + "step": 6704 + }, + { + "epoch": 93.12576419213974, + "grad_norm": 0.015600379556417465, + "learning_rate": 0.0006, + "loss": 4.0860700607299805, + "step": 6705 + }, + { + "epoch": 93.13973799126637, + "grad_norm": 0.017514333128929138, + "learning_rate": 0.0006, + "loss": 4.240778923034668, + "step": 6706 + }, + { + "epoch": 93.15371179039302, + "grad_norm": 0.01780967228114605, + "learning_rate": 0.0006, + "loss": 4.076499938964844, + "step": 6707 + }, + { + "epoch": 93.16768558951965, + "grad_norm": 0.016370438039302826, + "learning_rate": 0.0006, + "loss": 4.0677618980407715, + "step": 6708 + }, + { + "epoch": 93.18165938864628, + "grad_norm": 0.015205491334199905, + "learning_rate": 0.0006, + "loss": 4.004945755004883, + "step": 6709 + }, + { + "epoch": 93.19563318777293, + "grad_norm": 0.01537297572940588, + "learning_rate": 0.0006, + "loss": 4.121147155761719, + "step": 6710 + }, + { + "epoch": 93.20960698689956, + "grad_norm": 0.017343711107969284, + "learning_rate": 0.0006, + "loss": 4.103215217590332, + "step": 6711 + }, + { + "epoch": 93.2235807860262, + "grad_norm": 0.01618131436407566, + "learning_rate": 0.0006, + "loss": 4.083741664886475, + "step": 6712 + }, + { + "epoch": 93.23755458515284, + "grad_norm": 0.013340235687792301, + "learning_rate": 0.0006, + "loss": 4.055032730102539, + "step": 6713 + }, + { + "epoch": 93.25152838427948, + "grad_norm": 0.014084120281040668, + "learning_rate": 0.0006, + "loss": 4.005610942840576, + "step": 6714 + }, + { + "epoch": 93.26550218340611, + "grad_norm": 0.016104530543088913, + "learning_rate": 0.0006, + "loss": 4.091764450073242, + "step": 6715 + }, + { + "epoch": 93.27947598253274, + "grad_norm": 0.015232563950121403, + "learning_rate": 0.0006, + "loss": 4.105735778808594, + "step": 6716 + }, + { + "epoch": 93.29344978165939, + "grad_norm": 0.014003508724272251, + "learning_rate": 0.0006, + "loss": 4.05783224105835, + "step": 6717 + }, + { + "epoch": 93.30742358078602, + "grad_norm": 0.01614833064377308, + "learning_rate": 0.0006, + "loss": 4.124567031860352, + "step": 6718 + }, + { + "epoch": 93.32139737991267, + "grad_norm": 0.015010425820946693, + "learning_rate": 0.0006, + "loss": 4.07656192779541, + "step": 6719 + }, + { + "epoch": 93.3353711790393, + "grad_norm": 0.01444012951105833, + "learning_rate": 0.0006, + "loss": 4.122415542602539, + "step": 6720 + }, + { + "epoch": 93.34934497816595, + "grad_norm": 0.015715785324573517, + "learning_rate": 0.0006, + "loss": 4.070283889770508, + "step": 6721 + }, + { + "epoch": 93.36331877729258, + "grad_norm": 0.014976450242102146, + "learning_rate": 0.0006, + "loss": 4.087043762207031, + "step": 6722 + }, + { + "epoch": 93.37729257641921, + "grad_norm": 0.014868560247123241, + "learning_rate": 0.0006, + "loss": 3.9671802520751953, + "step": 6723 + }, + { + "epoch": 93.39126637554585, + "grad_norm": 0.013518013060092926, + "learning_rate": 0.0006, + "loss": 4.136781692504883, + "step": 6724 + }, + { + "epoch": 93.40524017467249, + "grad_norm": 0.013498986139893532, + "learning_rate": 0.0006, + "loss": 3.982309103012085, + "step": 6725 + }, + { + "epoch": 93.41921397379913, + "grad_norm": 0.013707910664379597, + "learning_rate": 0.0006, + "loss": 4.181885719299316, + "step": 6726 + }, + { + "epoch": 93.43318777292576, + "grad_norm": 0.013913207687437534, + "learning_rate": 0.0006, + "loss": 3.980070114135742, + "step": 6727 + }, + { + "epoch": 93.44716157205241, + "grad_norm": 0.013169893994927406, + "learning_rate": 0.0006, + "loss": 4.023380279541016, + "step": 6728 + }, + { + "epoch": 93.46113537117904, + "grad_norm": 0.014451324008405209, + "learning_rate": 0.0006, + "loss": 4.038372039794922, + "step": 6729 + }, + { + "epoch": 93.47510917030567, + "grad_norm": 0.014401541091501713, + "learning_rate": 0.0006, + "loss": 4.069009304046631, + "step": 6730 + }, + { + "epoch": 93.48908296943232, + "grad_norm": 0.013971361331641674, + "learning_rate": 0.0006, + "loss": 4.105923652648926, + "step": 6731 + }, + { + "epoch": 93.50305676855895, + "grad_norm": 0.013192536309361458, + "learning_rate": 0.0006, + "loss": 4.07828950881958, + "step": 6732 + }, + { + "epoch": 93.5170305676856, + "grad_norm": 0.013581741601228714, + "learning_rate": 0.0006, + "loss": 4.079203128814697, + "step": 6733 + }, + { + "epoch": 93.53100436681223, + "grad_norm": 0.01508194301277399, + "learning_rate": 0.0006, + "loss": 4.084749221801758, + "step": 6734 + }, + { + "epoch": 93.54497816593886, + "grad_norm": 0.0156723503023386, + "learning_rate": 0.0006, + "loss": 4.070216178894043, + "step": 6735 + }, + { + "epoch": 93.5589519650655, + "grad_norm": 0.014996538870036602, + "learning_rate": 0.0006, + "loss": 4.111318588256836, + "step": 6736 + }, + { + "epoch": 93.57292576419214, + "grad_norm": 0.012815269641578197, + "learning_rate": 0.0006, + "loss": 4.130414962768555, + "step": 6737 + }, + { + "epoch": 93.58689956331878, + "grad_norm": 0.012887579388916492, + "learning_rate": 0.0006, + "loss": 4.128446102142334, + "step": 6738 + }, + { + "epoch": 93.60087336244541, + "grad_norm": 0.013632671907544136, + "learning_rate": 0.0006, + "loss": 4.20977258682251, + "step": 6739 + }, + { + "epoch": 93.61484716157206, + "grad_norm": 0.014318895526230335, + "learning_rate": 0.0006, + "loss": 4.005739212036133, + "step": 6740 + }, + { + "epoch": 93.62882096069869, + "grad_norm": 0.014160127379000187, + "learning_rate": 0.0006, + "loss": 4.122544288635254, + "step": 6741 + }, + { + "epoch": 93.64279475982532, + "grad_norm": 0.01308423187583685, + "learning_rate": 0.0006, + "loss": 4.108774662017822, + "step": 6742 + }, + { + "epoch": 93.65676855895197, + "grad_norm": 0.014293737709522247, + "learning_rate": 0.0006, + "loss": 3.983156681060791, + "step": 6743 + }, + { + "epoch": 93.6707423580786, + "grad_norm": 0.015155135653913021, + "learning_rate": 0.0006, + "loss": 4.006117820739746, + "step": 6744 + }, + { + "epoch": 93.68471615720524, + "grad_norm": 0.014391839504241943, + "learning_rate": 0.0006, + "loss": 4.1658935546875, + "step": 6745 + }, + { + "epoch": 93.69868995633188, + "grad_norm": 0.014924845658242702, + "learning_rate": 0.0006, + "loss": 4.090868949890137, + "step": 6746 + }, + { + "epoch": 93.71266375545852, + "grad_norm": 0.014786627143621445, + "learning_rate": 0.0006, + "loss": 4.121716499328613, + "step": 6747 + }, + { + "epoch": 93.72663755458515, + "grad_norm": 0.016948824748396873, + "learning_rate": 0.0006, + "loss": 4.213179111480713, + "step": 6748 + }, + { + "epoch": 93.74061135371178, + "grad_norm": 0.01732620596885681, + "learning_rate": 0.0006, + "loss": 4.079215049743652, + "step": 6749 + }, + { + "epoch": 93.75458515283843, + "grad_norm": 0.016298124566674232, + "learning_rate": 0.0006, + "loss": 4.096126079559326, + "step": 6750 + }, + { + "epoch": 93.76855895196506, + "grad_norm": 0.015744732692837715, + "learning_rate": 0.0006, + "loss": 4.046300411224365, + "step": 6751 + }, + { + "epoch": 93.78253275109171, + "grad_norm": 0.016330812126398087, + "learning_rate": 0.0006, + "loss": 4.098011016845703, + "step": 6752 + }, + { + "epoch": 93.79650655021834, + "grad_norm": 0.01710600219666958, + "learning_rate": 0.0006, + "loss": 4.031402587890625, + "step": 6753 + }, + { + "epoch": 93.81048034934499, + "grad_norm": 0.01561805047094822, + "learning_rate": 0.0006, + "loss": 4.023388862609863, + "step": 6754 + }, + { + "epoch": 93.82445414847162, + "grad_norm": 0.013637243770062923, + "learning_rate": 0.0006, + "loss": 4.143238067626953, + "step": 6755 + }, + { + "epoch": 93.83842794759825, + "grad_norm": 0.013088075444102287, + "learning_rate": 0.0006, + "loss": 3.9083704948425293, + "step": 6756 + }, + { + "epoch": 93.8524017467249, + "grad_norm": 0.015893537551164627, + "learning_rate": 0.0006, + "loss": 4.102263450622559, + "step": 6757 + }, + { + "epoch": 93.86637554585153, + "grad_norm": 0.015213928185403347, + "learning_rate": 0.0006, + "loss": 4.111978530883789, + "step": 6758 + }, + { + "epoch": 93.88034934497817, + "grad_norm": 0.014543982222676277, + "learning_rate": 0.0006, + "loss": 4.05889892578125, + "step": 6759 + }, + { + "epoch": 93.8943231441048, + "grad_norm": 0.015104752965271473, + "learning_rate": 0.0006, + "loss": 4.136166572570801, + "step": 6760 + }, + { + "epoch": 93.90829694323143, + "grad_norm": 0.01505881268531084, + "learning_rate": 0.0006, + "loss": 4.061883449554443, + "step": 6761 + }, + { + "epoch": 93.92227074235808, + "grad_norm": 0.015248455107212067, + "learning_rate": 0.0006, + "loss": 4.109863758087158, + "step": 6762 + }, + { + "epoch": 93.93624454148471, + "grad_norm": 0.013651185669004917, + "learning_rate": 0.0006, + "loss": 4.061549663543701, + "step": 6763 + }, + { + "epoch": 93.95021834061136, + "grad_norm": 0.013806282542645931, + "learning_rate": 0.0006, + "loss": 4.205403804779053, + "step": 6764 + }, + { + "epoch": 93.96419213973799, + "grad_norm": 0.01538451761007309, + "learning_rate": 0.0006, + "loss": 4.14816951751709, + "step": 6765 + }, + { + "epoch": 93.97816593886463, + "grad_norm": 0.01602427288889885, + "learning_rate": 0.0006, + "loss": 4.000303268432617, + "step": 6766 + }, + { + "epoch": 93.99213973799127, + "grad_norm": 0.014558793976902962, + "learning_rate": 0.0006, + "loss": 4.096562385559082, + "step": 6767 + }, + { + "epoch": 94.0, + "grad_norm": 0.01613018475472927, + "learning_rate": 0.0006, + "loss": 4.262835502624512, + "step": 6768 + }, + { + "epoch": 94.0, + "eval_loss": 4.532126426696777, + "eval_runtime": 56.7592, + "eval_samples_per_second": 43.024, + "eval_steps_per_second": 1.357, + "step": 6768 + }, + { + "epoch": 94.01397379912663, + "grad_norm": 0.01614535041153431, + "learning_rate": 0.0006, + "loss": 4.073431491851807, + "step": 6769 + }, + { + "epoch": 94.02794759825328, + "grad_norm": 0.015986818820238113, + "learning_rate": 0.0006, + "loss": 4.070001602172852, + "step": 6770 + }, + { + "epoch": 94.04192139737991, + "grad_norm": 0.015678714960813522, + "learning_rate": 0.0006, + "loss": 4.06920051574707, + "step": 6771 + }, + { + "epoch": 94.05589519650655, + "grad_norm": 0.016593851149082184, + "learning_rate": 0.0006, + "loss": 4.095780849456787, + "step": 6772 + }, + { + "epoch": 94.06986899563319, + "grad_norm": 0.01780948042869568, + "learning_rate": 0.0006, + "loss": 4.092916488647461, + "step": 6773 + }, + { + "epoch": 94.08384279475983, + "grad_norm": 0.01755528897047043, + "learning_rate": 0.0006, + "loss": 3.9833455085754395, + "step": 6774 + }, + { + "epoch": 94.09781659388646, + "grad_norm": 0.01708311215043068, + "learning_rate": 0.0006, + "loss": 4.053336143493652, + "step": 6775 + }, + { + "epoch": 94.1117903930131, + "grad_norm": 0.017362922430038452, + "learning_rate": 0.0006, + "loss": 4.014343738555908, + "step": 6776 + }, + { + "epoch": 94.12576419213974, + "grad_norm": 0.01824064739048481, + "learning_rate": 0.0006, + "loss": 4.1215691566467285, + "step": 6777 + }, + { + "epoch": 94.13973799126637, + "grad_norm": 0.017743906006217003, + "learning_rate": 0.0006, + "loss": 4.091215133666992, + "step": 6778 + }, + { + "epoch": 94.15371179039302, + "grad_norm": 0.016585713252425194, + "learning_rate": 0.0006, + "loss": 4.013662338256836, + "step": 6779 + }, + { + "epoch": 94.16768558951965, + "grad_norm": 0.017397213727235794, + "learning_rate": 0.0006, + "loss": 4.070602893829346, + "step": 6780 + }, + { + "epoch": 94.18165938864628, + "grad_norm": 0.01733497530221939, + "learning_rate": 0.0006, + "loss": 4.1304426193237305, + "step": 6781 + }, + { + "epoch": 94.19563318777293, + "grad_norm": 0.015215300023555756, + "learning_rate": 0.0006, + "loss": 3.965014696121216, + "step": 6782 + }, + { + "epoch": 94.20960698689956, + "grad_norm": 0.015908699482679367, + "learning_rate": 0.0006, + "loss": 3.924964427947998, + "step": 6783 + }, + { + "epoch": 94.2235807860262, + "grad_norm": 0.014044968411326408, + "learning_rate": 0.0006, + "loss": 4.08673095703125, + "step": 6784 + }, + { + "epoch": 94.23755458515284, + "grad_norm": 0.01375142764300108, + "learning_rate": 0.0006, + "loss": 4.011466979980469, + "step": 6785 + }, + { + "epoch": 94.25152838427948, + "grad_norm": 0.015539257787168026, + "learning_rate": 0.0006, + "loss": 4.09976863861084, + "step": 6786 + }, + { + "epoch": 94.26550218340611, + "grad_norm": 0.014303367584943771, + "learning_rate": 0.0006, + "loss": 4.108369827270508, + "step": 6787 + }, + { + "epoch": 94.27947598253274, + "grad_norm": 0.01281928364187479, + "learning_rate": 0.0006, + "loss": 4.073737144470215, + "step": 6788 + }, + { + "epoch": 94.29344978165939, + "grad_norm": 0.013859959319233894, + "learning_rate": 0.0006, + "loss": 4.054823875427246, + "step": 6789 + }, + { + "epoch": 94.30742358078602, + "grad_norm": 0.013326860964298248, + "learning_rate": 0.0006, + "loss": 4.133251190185547, + "step": 6790 + }, + { + "epoch": 94.32139737991267, + "grad_norm": 0.013923076912760735, + "learning_rate": 0.0006, + "loss": 4.190706729888916, + "step": 6791 + }, + { + "epoch": 94.3353711790393, + "grad_norm": 0.01420847699046135, + "learning_rate": 0.0006, + "loss": 3.9336605072021484, + "step": 6792 + }, + { + "epoch": 94.34934497816595, + "grad_norm": 0.014607555232942104, + "learning_rate": 0.0006, + "loss": 4.128399848937988, + "step": 6793 + }, + { + "epoch": 94.36331877729258, + "grad_norm": 0.012862961739301682, + "learning_rate": 0.0006, + "loss": 4.0132527351379395, + "step": 6794 + }, + { + "epoch": 94.37729257641921, + "grad_norm": 0.012530217878520489, + "learning_rate": 0.0006, + "loss": 4.0384368896484375, + "step": 6795 + }, + { + "epoch": 94.39126637554585, + "grad_norm": 0.01286802627146244, + "learning_rate": 0.0006, + "loss": 3.95570707321167, + "step": 6796 + }, + { + "epoch": 94.40524017467249, + "grad_norm": 0.013450917787849903, + "learning_rate": 0.0006, + "loss": 4.016484260559082, + "step": 6797 + }, + { + "epoch": 94.41921397379913, + "grad_norm": 0.01393423043191433, + "learning_rate": 0.0006, + "loss": 4.124959468841553, + "step": 6798 + }, + { + "epoch": 94.43318777292576, + "grad_norm": 0.013077184557914734, + "learning_rate": 0.0006, + "loss": 4.029163837432861, + "step": 6799 + }, + { + "epoch": 94.44716157205241, + "grad_norm": 0.013577334582805634, + "learning_rate": 0.0006, + "loss": 4.22929573059082, + "step": 6800 + }, + { + "epoch": 94.46113537117904, + "grad_norm": 0.013930312357842922, + "learning_rate": 0.0006, + "loss": 4.024711608886719, + "step": 6801 + }, + { + "epoch": 94.47510917030567, + "grad_norm": 0.013746250420808792, + "learning_rate": 0.0006, + "loss": 3.9766428470611572, + "step": 6802 + }, + { + "epoch": 94.48908296943232, + "grad_norm": 0.013682983815670013, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6803 + }, + { + "epoch": 94.50305676855895, + "grad_norm": 0.01473842654377222, + "learning_rate": 0.0006, + "loss": 4.197623252868652, + "step": 6804 + }, + { + "epoch": 94.5170305676856, + "grad_norm": 0.014133758842945099, + "learning_rate": 0.0006, + "loss": 4.055549621582031, + "step": 6805 + }, + { + "epoch": 94.53100436681223, + "grad_norm": 0.0139645179733634, + "learning_rate": 0.0006, + "loss": 4.130135536193848, + "step": 6806 + }, + { + "epoch": 94.54497816593886, + "grad_norm": 0.013973365537822247, + "learning_rate": 0.0006, + "loss": 4.006230354309082, + "step": 6807 + }, + { + "epoch": 94.5589519650655, + "grad_norm": 0.014227126725018024, + "learning_rate": 0.0006, + "loss": 4.130459785461426, + "step": 6808 + }, + { + "epoch": 94.57292576419214, + "grad_norm": 0.013129506260156631, + "learning_rate": 0.0006, + "loss": 4.014974594116211, + "step": 6809 + }, + { + "epoch": 94.58689956331878, + "grad_norm": 0.013015891425311565, + "learning_rate": 0.0006, + "loss": 4.161370277404785, + "step": 6810 + }, + { + "epoch": 94.60087336244541, + "grad_norm": 0.013957147486507893, + "learning_rate": 0.0006, + "loss": 4.09519100189209, + "step": 6811 + }, + { + "epoch": 94.61484716157206, + "grad_norm": 0.015494572930037975, + "learning_rate": 0.0006, + "loss": 4.029949188232422, + "step": 6812 + }, + { + "epoch": 94.62882096069869, + "grad_norm": 0.016732869669795036, + "learning_rate": 0.0006, + "loss": 4.135669708251953, + "step": 6813 + }, + { + "epoch": 94.64279475982532, + "grad_norm": 0.01769322343170643, + "learning_rate": 0.0006, + "loss": 4.109872817993164, + "step": 6814 + }, + { + "epoch": 94.65676855895197, + "grad_norm": 0.015932537615299225, + "learning_rate": 0.0006, + "loss": 4.014595031738281, + "step": 6815 + }, + { + "epoch": 94.6707423580786, + "grad_norm": 0.015152990818023682, + "learning_rate": 0.0006, + "loss": 4.1983442306518555, + "step": 6816 + }, + { + "epoch": 94.68471615720524, + "grad_norm": 0.01590685173869133, + "learning_rate": 0.0006, + "loss": 4.119250774383545, + "step": 6817 + }, + { + "epoch": 94.69868995633188, + "grad_norm": 0.016018465161323547, + "learning_rate": 0.0006, + "loss": 4.048069953918457, + "step": 6818 + }, + { + "epoch": 94.71266375545852, + "grad_norm": 0.016621652990579605, + "learning_rate": 0.0006, + "loss": 4.043731689453125, + "step": 6819 + }, + { + "epoch": 94.72663755458515, + "grad_norm": 0.014535024762153625, + "learning_rate": 0.0006, + "loss": 4.1886820793151855, + "step": 6820 + }, + { + "epoch": 94.74061135371178, + "grad_norm": 0.014192107133567333, + "learning_rate": 0.0006, + "loss": 4.103938102722168, + "step": 6821 + }, + { + "epoch": 94.75458515283843, + "grad_norm": 0.0166022852063179, + "learning_rate": 0.0006, + "loss": 4.017818450927734, + "step": 6822 + }, + { + "epoch": 94.76855895196506, + "grad_norm": 0.016247155144810677, + "learning_rate": 0.0006, + "loss": 4.0038981437683105, + "step": 6823 + }, + { + "epoch": 94.78253275109171, + "grad_norm": 0.01593470573425293, + "learning_rate": 0.0006, + "loss": 4.043969631195068, + "step": 6824 + }, + { + "epoch": 94.79650655021834, + "grad_norm": 0.014334345236420631, + "learning_rate": 0.0006, + "loss": 4.097829341888428, + "step": 6825 + }, + { + "epoch": 94.81048034934499, + "grad_norm": 0.015581986866891384, + "learning_rate": 0.0006, + "loss": 4.148984432220459, + "step": 6826 + }, + { + "epoch": 94.82445414847162, + "grad_norm": 0.017519738525152206, + "learning_rate": 0.0006, + "loss": 4.245623588562012, + "step": 6827 + }, + { + "epoch": 94.83842794759825, + "grad_norm": 0.015155277214944363, + "learning_rate": 0.0006, + "loss": 4.031019687652588, + "step": 6828 + }, + { + "epoch": 94.8524017467249, + "grad_norm": 0.014969222247600555, + "learning_rate": 0.0006, + "loss": 4.036084175109863, + "step": 6829 + }, + { + "epoch": 94.86637554585153, + "grad_norm": 0.015987908467650414, + "learning_rate": 0.0006, + "loss": 4.08396053314209, + "step": 6830 + }, + { + "epoch": 94.88034934497817, + "grad_norm": 0.016324151307344437, + "learning_rate": 0.0006, + "loss": 4.036787033081055, + "step": 6831 + }, + { + "epoch": 94.8943231441048, + "grad_norm": 0.015957487747073174, + "learning_rate": 0.0006, + "loss": 4.147607803344727, + "step": 6832 + }, + { + "epoch": 94.90829694323143, + "grad_norm": 0.015286967158317566, + "learning_rate": 0.0006, + "loss": 4.184615135192871, + "step": 6833 + }, + { + "epoch": 94.92227074235808, + "grad_norm": 0.01338377594947815, + "learning_rate": 0.0006, + "loss": 4.014981269836426, + "step": 6834 + }, + { + "epoch": 94.93624454148471, + "grad_norm": 0.013997633941471577, + "learning_rate": 0.0006, + "loss": 4.021933078765869, + "step": 6835 + }, + { + "epoch": 94.95021834061136, + "grad_norm": 0.014572402462363243, + "learning_rate": 0.0006, + "loss": 4.0326948165893555, + "step": 6836 + }, + { + "epoch": 94.96419213973799, + "grad_norm": 0.014379586093127728, + "learning_rate": 0.0006, + "loss": 4.060697555541992, + "step": 6837 + }, + { + "epoch": 94.97816593886463, + "grad_norm": 0.013513042591512203, + "learning_rate": 0.0006, + "loss": 4.05123233795166, + "step": 6838 + }, + { + "epoch": 94.99213973799127, + "grad_norm": 0.012904108501970768, + "learning_rate": 0.0006, + "loss": 4.020852565765381, + "step": 6839 + }, + { + "epoch": 95.0, + "grad_norm": 0.014364209026098251, + "learning_rate": 0.0006, + "loss": 4.080837249755859, + "step": 6840 + }, + { + "epoch": 95.0, + "eval_loss": 4.608776569366455, + "eval_runtime": 56.6516, + "eval_samples_per_second": 43.106, + "eval_steps_per_second": 1.359, + "step": 6840 + }, + { + "epoch": 95.01397379912663, + "grad_norm": 0.01357510220259428, + "learning_rate": 0.0006, + "loss": 4.105312347412109, + "step": 6841 + }, + { + "epoch": 95.02794759825328, + "grad_norm": 0.014248257502913475, + "learning_rate": 0.0006, + "loss": 3.971667766571045, + "step": 6842 + }, + { + "epoch": 95.04192139737991, + "grad_norm": 0.014773483388125896, + "learning_rate": 0.0006, + "loss": 4.03553581237793, + "step": 6843 + }, + { + "epoch": 95.05589519650655, + "grad_norm": 0.014766111969947815, + "learning_rate": 0.0006, + "loss": 4.074707508087158, + "step": 6844 + }, + { + "epoch": 95.06986899563319, + "grad_norm": 0.014778018929064274, + "learning_rate": 0.0006, + "loss": 4.1083855628967285, + "step": 6845 + }, + { + "epoch": 95.08384279475983, + "grad_norm": 0.014466686174273491, + "learning_rate": 0.0006, + "loss": 4.054587364196777, + "step": 6846 + }, + { + "epoch": 95.09781659388646, + "grad_norm": 0.015195904299616814, + "learning_rate": 0.0006, + "loss": 4.142580032348633, + "step": 6847 + }, + { + "epoch": 95.1117903930131, + "grad_norm": 0.01530557032674551, + "learning_rate": 0.0006, + "loss": 4.062134265899658, + "step": 6848 + }, + { + "epoch": 95.12576419213974, + "grad_norm": 0.015203768387436867, + "learning_rate": 0.0006, + "loss": 4.082962989807129, + "step": 6849 + }, + { + "epoch": 95.13973799126637, + "grad_norm": 0.014906628057360649, + "learning_rate": 0.0006, + "loss": 4.033127784729004, + "step": 6850 + }, + { + "epoch": 95.15371179039302, + "grad_norm": 0.01553534809499979, + "learning_rate": 0.0006, + "loss": 4.051743030548096, + "step": 6851 + }, + { + "epoch": 95.16768558951965, + "grad_norm": 0.01607976108789444, + "learning_rate": 0.0006, + "loss": 4.038073539733887, + "step": 6852 + }, + { + "epoch": 95.18165938864628, + "grad_norm": 0.015776360407471657, + "learning_rate": 0.0006, + "loss": 4.013091087341309, + "step": 6853 + }, + { + "epoch": 95.19563318777293, + "grad_norm": 0.01623626798391342, + "learning_rate": 0.0006, + "loss": 4.12293815612793, + "step": 6854 + }, + { + "epoch": 95.20960698689956, + "grad_norm": 0.016887828707695007, + "learning_rate": 0.0006, + "loss": 4.120452404022217, + "step": 6855 + }, + { + "epoch": 95.2235807860262, + "grad_norm": 0.01561128068715334, + "learning_rate": 0.0006, + "loss": 4.071322441101074, + "step": 6856 + }, + { + "epoch": 95.23755458515284, + "grad_norm": 0.015679724514484406, + "learning_rate": 0.0006, + "loss": 4.065072059631348, + "step": 6857 + }, + { + "epoch": 95.25152838427948, + "grad_norm": 0.016506962478160858, + "learning_rate": 0.0006, + "loss": 4.0667924880981445, + "step": 6858 + }, + { + "epoch": 95.26550218340611, + "grad_norm": 0.015340839512646198, + "learning_rate": 0.0006, + "loss": 4.14047908782959, + "step": 6859 + }, + { + "epoch": 95.27947598253274, + "grad_norm": 0.013935413211584091, + "learning_rate": 0.0006, + "loss": 3.9688191413879395, + "step": 6860 + }, + { + "epoch": 95.29344978165939, + "grad_norm": 0.015223405323922634, + "learning_rate": 0.0006, + "loss": 4.101613521575928, + "step": 6861 + }, + { + "epoch": 95.30742358078602, + "grad_norm": 0.01562546379864216, + "learning_rate": 0.0006, + "loss": 4.128626823425293, + "step": 6862 + }, + { + "epoch": 95.32139737991267, + "grad_norm": 0.016159312799572945, + "learning_rate": 0.0006, + "loss": 4.080667018890381, + "step": 6863 + }, + { + "epoch": 95.3353711790393, + "grad_norm": 0.01724511943757534, + "learning_rate": 0.0006, + "loss": 4.08463191986084, + "step": 6864 + }, + { + "epoch": 95.34934497816595, + "grad_norm": 0.01592276245355606, + "learning_rate": 0.0006, + "loss": 4.127612113952637, + "step": 6865 + }, + { + "epoch": 95.36331877729258, + "grad_norm": 0.01744082197546959, + "learning_rate": 0.0006, + "loss": 4.034480094909668, + "step": 6866 + }, + { + "epoch": 95.37729257641921, + "grad_norm": 0.01774069108068943, + "learning_rate": 0.0006, + "loss": 4.091196537017822, + "step": 6867 + }, + { + "epoch": 95.39126637554585, + "grad_norm": 0.01653675176203251, + "learning_rate": 0.0006, + "loss": 4.032937526702881, + "step": 6868 + }, + { + "epoch": 95.40524017467249, + "grad_norm": 0.014933439902961254, + "learning_rate": 0.0006, + "loss": 3.9386508464813232, + "step": 6869 + }, + { + "epoch": 95.41921397379913, + "grad_norm": 0.014743257313966751, + "learning_rate": 0.0006, + "loss": 4.126462459564209, + "step": 6870 + }, + { + "epoch": 95.43318777292576, + "grad_norm": 0.014614643529057503, + "learning_rate": 0.0006, + "loss": 4.093838214874268, + "step": 6871 + }, + { + "epoch": 95.44716157205241, + "grad_norm": 0.01665084809064865, + "learning_rate": 0.0006, + "loss": 4.1471967697143555, + "step": 6872 + }, + { + "epoch": 95.46113537117904, + "grad_norm": 0.018056001514196396, + "learning_rate": 0.0006, + "loss": 4.1388349533081055, + "step": 6873 + }, + { + "epoch": 95.47510917030567, + "grad_norm": 0.02129282057285309, + "learning_rate": 0.0006, + "loss": 4.130726337432861, + "step": 6874 + }, + { + "epoch": 95.48908296943232, + "grad_norm": 0.022237276658415794, + "learning_rate": 0.0006, + "loss": 4.10750675201416, + "step": 6875 + }, + { + "epoch": 95.50305676855895, + "grad_norm": 0.021106960251927376, + "learning_rate": 0.0006, + "loss": 4.136451721191406, + "step": 6876 + }, + { + "epoch": 95.5170305676856, + "grad_norm": 0.019107872620224953, + "learning_rate": 0.0006, + "loss": 4.139644145965576, + "step": 6877 + }, + { + "epoch": 95.53100436681223, + "grad_norm": 0.018290145322680473, + "learning_rate": 0.0006, + "loss": 4.161884784698486, + "step": 6878 + }, + { + "epoch": 95.54497816593886, + "grad_norm": 0.018161306157708168, + "learning_rate": 0.0006, + "loss": 4.1252336502075195, + "step": 6879 + }, + { + "epoch": 95.5589519650655, + "grad_norm": 0.018432756885886192, + "learning_rate": 0.0006, + "loss": 4.0635986328125, + "step": 6880 + }, + { + "epoch": 95.57292576419214, + "grad_norm": 0.016592558473348618, + "learning_rate": 0.0006, + "loss": 4.037112236022949, + "step": 6881 + }, + { + "epoch": 95.58689956331878, + "grad_norm": 0.015283452346920967, + "learning_rate": 0.0006, + "loss": 4.1964826583862305, + "step": 6882 + }, + { + "epoch": 95.60087336244541, + "grad_norm": 0.015539180487394333, + "learning_rate": 0.0006, + "loss": 4.110917568206787, + "step": 6883 + }, + { + "epoch": 95.61484716157206, + "grad_norm": 0.016508113592863083, + "learning_rate": 0.0006, + "loss": 3.9705517292022705, + "step": 6884 + }, + { + "epoch": 95.62882096069869, + "grad_norm": 0.016635406762361526, + "learning_rate": 0.0006, + "loss": 4.025636196136475, + "step": 6885 + }, + { + "epoch": 95.64279475982532, + "grad_norm": 0.01688900962471962, + "learning_rate": 0.0006, + "loss": 4.01814079284668, + "step": 6886 + }, + { + "epoch": 95.65676855895197, + "grad_norm": 0.01614062674343586, + "learning_rate": 0.0006, + "loss": 4.098041534423828, + "step": 6887 + }, + { + "epoch": 95.6707423580786, + "grad_norm": 0.015679053962230682, + "learning_rate": 0.0006, + "loss": 4.077049255371094, + "step": 6888 + }, + { + "epoch": 95.68471615720524, + "grad_norm": 0.016617605462670326, + "learning_rate": 0.0006, + "loss": 4.10363245010376, + "step": 6889 + }, + { + "epoch": 95.69868995633188, + "grad_norm": 0.015966709703207016, + "learning_rate": 0.0006, + "loss": 4.100924491882324, + "step": 6890 + }, + { + "epoch": 95.71266375545852, + "grad_norm": 0.015768567100167274, + "learning_rate": 0.0006, + "loss": 4.063976287841797, + "step": 6891 + }, + { + "epoch": 95.72663755458515, + "grad_norm": 0.015800341963768005, + "learning_rate": 0.0006, + "loss": 4.008409023284912, + "step": 6892 + }, + { + "epoch": 95.74061135371178, + "grad_norm": 0.014288571663200855, + "learning_rate": 0.0006, + "loss": 4.139591217041016, + "step": 6893 + }, + { + "epoch": 95.75458515283843, + "grad_norm": 0.013743448071181774, + "learning_rate": 0.0006, + "loss": 3.967352867126465, + "step": 6894 + }, + { + "epoch": 95.76855895196506, + "grad_norm": 0.01387989055365324, + "learning_rate": 0.0006, + "loss": 4.127554893493652, + "step": 6895 + }, + { + "epoch": 95.78253275109171, + "grad_norm": 0.013910802081227303, + "learning_rate": 0.0006, + "loss": 4.130941390991211, + "step": 6896 + }, + { + "epoch": 95.79650655021834, + "grad_norm": 0.013936805538833141, + "learning_rate": 0.0006, + "loss": 4.138744831085205, + "step": 6897 + }, + { + "epoch": 95.81048034934499, + "grad_norm": 0.015135723166167736, + "learning_rate": 0.0006, + "loss": 4.131860256195068, + "step": 6898 + }, + { + "epoch": 95.82445414847162, + "grad_norm": 0.014921358786523342, + "learning_rate": 0.0006, + "loss": 4.1949462890625, + "step": 6899 + }, + { + "epoch": 95.83842794759825, + "grad_norm": 0.01401488296687603, + "learning_rate": 0.0006, + "loss": 4.181267738342285, + "step": 6900 + }, + { + "epoch": 95.8524017467249, + "grad_norm": 0.013691626489162445, + "learning_rate": 0.0006, + "loss": 4.113667964935303, + "step": 6901 + }, + { + "epoch": 95.86637554585153, + "grad_norm": 0.013890047557651997, + "learning_rate": 0.0006, + "loss": 4.163067817687988, + "step": 6902 + }, + { + "epoch": 95.88034934497817, + "grad_norm": 0.01385503076016903, + "learning_rate": 0.0006, + "loss": 4.076153755187988, + "step": 6903 + }, + { + "epoch": 95.8943231441048, + "grad_norm": 0.01383926346898079, + "learning_rate": 0.0006, + "loss": 4.102423667907715, + "step": 6904 + }, + { + "epoch": 95.90829694323143, + "grad_norm": 0.015961717814207077, + "learning_rate": 0.0006, + "loss": 4.1543779373168945, + "step": 6905 + }, + { + "epoch": 95.92227074235808, + "grad_norm": 0.016718363389372826, + "learning_rate": 0.0006, + "loss": 3.991607189178467, + "step": 6906 + }, + { + "epoch": 95.93624454148471, + "grad_norm": 0.01585792563855648, + "learning_rate": 0.0006, + "loss": 4.048587799072266, + "step": 6907 + }, + { + "epoch": 95.95021834061136, + "grad_norm": 0.015311713330447674, + "learning_rate": 0.0006, + "loss": 4.1070451736450195, + "step": 6908 + }, + { + "epoch": 95.96419213973799, + "grad_norm": 0.014521263539791107, + "learning_rate": 0.0006, + "loss": 4.087587356567383, + "step": 6909 + }, + { + "epoch": 95.97816593886463, + "grad_norm": 0.01433157455176115, + "learning_rate": 0.0006, + "loss": 4.139904499053955, + "step": 6910 + }, + { + "epoch": 95.99213973799127, + "grad_norm": 0.014475514180958271, + "learning_rate": 0.0006, + "loss": 4.182525634765625, + "step": 6911 + }, + { + "epoch": 96.0, + "grad_norm": 0.015245620161294937, + "learning_rate": 0.0006, + "loss": 4.1199870109558105, + "step": 6912 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.946828826878935e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-6912/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-6912/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-6912/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-72/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3cd56dfadca87c7ad5ef1d49ed08fdd77955d583 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e92299b482a10bf724c8131c23f573f479a18d94a8335b3d22a1c327e7b4922 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b7eeeb97cf14b488246532851ad85039b6f5910a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a049079dd03efbc2f0be2294231a6d522211af4dd67fad4631b8bbb7875df1c6 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..942953431b6b69250b3ac046cdf5cca437235031 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54c9019619b0417fc7c61a86a46149eb207b6bbd0181871dcec50cb454d3f367 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..ea26b0c2f5adce832b4d8762855e8be5382d63d7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4dddd0ae188208a24e5ab9a9fc1928ad2f9275ce147a0a992232c03bba455215 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..409136377918f95867748fce9457de5970be4a6f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8068dd390bcd57d32f1dfc5e792134a165f320ca31900b27252a73a53d7ec360 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..7eaf02283cfd4b5f41ba417c553b0f39b696450e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6014b1c6c1718e10a75cbe0f7295e4d1975ef1aa122296fb2c3b54f4b1835c45 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..eee91baef0b1848c0b5219cb5ba7b8365d12789d --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json @@ -0,0 +1,538 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 72, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.069613361332224e+16, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-72/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-72/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-72/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-7200/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-7200/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..11b1b00a9fb8630c6045e70e23bd9ee0ac04b991 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:992ea84d161c0ced7c124ed737b6e624e3398b7f1573ca2595bcdaf054443a17 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7200/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..cc7b60119cc8d73a8da7efe1ef589422eb496662 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acc9564263da743645154f940ed413297dcb67ed8abb28c36c62eb1619aebaa8 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..e5981255de81c81160f3351d04278830d3bb2507 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fc4fb416576a2293bd2b09eb738accb4d54bde927d92f2fe90be3a448c990e5 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..edd57664d15ac959bb114a9d87d65d0436eb2365 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69ae21eb3a416db05955df400b5252da17a3f027ae01a9ee9dcf0c5b62137ce1 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..fe7b2d78121a938380d42815769910f27018d664 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:907ef0007bd49a0897d16fbbeecd39ead8005845e40b3dffeb613c567c7d36e0 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..48e04b3f4c6cb13a09a4fc89cc4a3af943e22dc5 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bbf8a37dcf5f2f1715be25d2aae1d740d39d750efa40b777134b7c66ef0bff1 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7200/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..624c26e041c29e4eef4955201869827f6f6a567e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f84784070f59d3d6db22567a60409a64f07923bdf1c69e2acde1a156c945423 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-7200/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..6dad79039d68014c806eb6ddcc522fb63745620e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/trainer_state.json @@ -0,0 +1,51226 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 100.0, + "eval_steps": 500, + "global_step": 7200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + }, + { + "epoch": 76.0, + "eval_loss": 4.689640998840332, + "eval_runtime": 56.9095, + "eval_samples_per_second": 42.91, + "eval_steps_per_second": 1.353, + "step": 5472 + }, + { + "epoch": 76.01397379912663, + "grad_norm": 0.015351055189967155, + "learning_rate": 0.0006, + "loss": 4.348268985748291, + "step": 5473 + }, + { + "epoch": 76.02794759825328, + "grad_norm": 0.017164217308163643, + "learning_rate": 0.0006, + "loss": 4.244204998016357, + "step": 5474 + }, + { + "epoch": 76.04192139737991, + "grad_norm": 0.021624932065606117, + "learning_rate": 0.0006, + "loss": 4.311168670654297, + "step": 5475 + }, + { + "epoch": 76.05589519650655, + "grad_norm": 0.028712468221783638, + "learning_rate": 0.0006, + "loss": 4.322287082672119, + "step": 5476 + }, + { + "epoch": 76.06986899563319, + "grad_norm": 0.027915487065911293, + "learning_rate": 0.0006, + "loss": 4.381207466125488, + "step": 5477 + }, + { + "epoch": 76.08384279475983, + "grad_norm": 0.017955303192138672, + "learning_rate": 0.0006, + "loss": 4.2635016441345215, + "step": 5478 + }, + { + "epoch": 76.09781659388646, + "grad_norm": 0.019113514572381973, + "learning_rate": 0.0006, + "loss": 4.399471282958984, + "step": 5479 + }, + { + "epoch": 76.1117903930131, + "grad_norm": 0.017502542585134506, + "learning_rate": 0.0006, + "loss": 4.321204662322998, + "step": 5480 + }, + { + "epoch": 76.12576419213974, + "grad_norm": 0.01852525770664215, + "learning_rate": 0.0006, + "loss": 4.217884063720703, + "step": 5481 + }, + { + "epoch": 76.13973799126637, + "grad_norm": 0.019836828112602234, + "learning_rate": 0.0006, + "loss": 4.278532028198242, + "step": 5482 + }, + { + "epoch": 76.15371179039302, + "grad_norm": 0.02101975679397583, + "learning_rate": 0.0006, + "loss": 4.246720314025879, + "step": 5483 + }, + { + "epoch": 76.16768558951965, + "grad_norm": 0.021202990785241127, + "learning_rate": 0.0006, + "loss": 4.517517566680908, + "step": 5484 + }, + { + "epoch": 76.18165938864628, + "grad_norm": 0.021428769454360008, + "learning_rate": 0.0006, + "loss": 4.3373517990112305, + "step": 5485 + }, + { + "epoch": 76.19563318777293, + "grad_norm": 0.018263190984725952, + "learning_rate": 0.0006, + "loss": 4.402856826782227, + "step": 5486 + }, + { + "epoch": 76.20960698689956, + "grad_norm": 0.019025536254048347, + "learning_rate": 0.0006, + "loss": 4.314227104187012, + "step": 5487 + }, + { + "epoch": 76.2235807860262, + "grad_norm": 0.020918630063533783, + "learning_rate": 0.0006, + "loss": 4.454369068145752, + "step": 5488 + }, + { + "epoch": 76.23755458515284, + "grad_norm": 0.019307617098093033, + "learning_rate": 0.0006, + "loss": 4.247509002685547, + "step": 5489 + }, + { + "epoch": 76.25152838427948, + "grad_norm": 0.014956331811845303, + "learning_rate": 0.0006, + "loss": 4.365175247192383, + "step": 5490 + }, + { + "epoch": 76.26550218340611, + "grad_norm": 0.017213059589266777, + "learning_rate": 0.0006, + "loss": 4.384005069732666, + "step": 5491 + }, + { + "epoch": 76.27947598253274, + "grad_norm": 0.017575597390532494, + "learning_rate": 0.0006, + "loss": 4.41934871673584, + "step": 5492 + }, + { + "epoch": 76.29344978165939, + "grad_norm": 0.015879256650805473, + "learning_rate": 0.0006, + "loss": 4.26675271987915, + "step": 5493 + }, + { + "epoch": 76.30742358078602, + "grad_norm": 0.01564522460103035, + "learning_rate": 0.0006, + "loss": 4.277339935302734, + "step": 5494 + }, + { + "epoch": 76.32139737991267, + "grad_norm": 0.016414586454629898, + "learning_rate": 0.0006, + "loss": 4.4138593673706055, + "step": 5495 + }, + { + "epoch": 76.3353711790393, + "grad_norm": 0.014732337556779385, + "learning_rate": 0.0006, + "loss": 4.281006336212158, + "step": 5496 + }, + { + "epoch": 76.34934497816595, + "grad_norm": 0.015461455099284649, + "learning_rate": 0.0006, + "loss": 4.263422966003418, + "step": 5497 + }, + { + "epoch": 76.36331877729258, + "grad_norm": 0.014529814012348652, + "learning_rate": 0.0006, + "loss": 4.36165189743042, + "step": 5498 + }, + { + "epoch": 76.37729257641921, + "grad_norm": 0.01469483319669962, + "learning_rate": 0.0006, + "loss": 4.309642791748047, + "step": 5499 + }, + { + "epoch": 76.39126637554585, + "grad_norm": 0.016490574926137924, + "learning_rate": 0.0006, + "loss": 4.288150787353516, + "step": 5500 + }, + { + "epoch": 76.40524017467249, + "grad_norm": 0.014593565836548805, + "learning_rate": 0.0006, + "loss": 4.271614074707031, + "step": 5501 + }, + { + "epoch": 76.41921397379913, + "grad_norm": 0.012796535156667233, + "learning_rate": 0.0006, + "loss": 4.420042037963867, + "step": 5502 + }, + { + "epoch": 76.43318777292576, + "grad_norm": 0.016086025163531303, + "learning_rate": 0.0006, + "loss": 4.289943695068359, + "step": 5503 + }, + { + "epoch": 76.44716157205241, + "grad_norm": 0.015439470298588276, + "learning_rate": 0.0006, + "loss": 4.3023271560668945, + "step": 5504 + }, + { + "epoch": 76.46113537117904, + "grad_norm": 0.015066367574036121, + "learning_rate": 0.0006, + "loss": 4.197081565856934, + "step": 5505 + }, + { + "epoch": 76.47510917030567, + "grad_norm": 0.016653263941407204, + "learning_rate": 0.0006, + "loss": 4.333785057067871, + "step": 5506 + }, + { + "epoch": 76.48908296943232, + "grad_norm": 0.014791973866522312, + "learning_rate": 0.0006, + "loss": 4.323811054229736, + "step": 5507 + }, + { + "epoch": 76.50305676855895, + "grad_norm": 0.014694827608764172, + "learning_rate": 0.0006, + "loss": 4.195852279663086, + "step": 5508 + }, + { + "epoch": 76.5170305676856, + "grad_norm": 0.015120396390557289, + "learning_rate": 0.0006, + "loss": 4.42379903793335, + "step": 5509 + }, + { + "epoch": 76.53100436681223, + "grad_norm": 0.014584025368094444, + "learning_rate": 0.0006, + "loss": 4.2702531814575195, + "step": 5510 + }, + { + "epoch": 76.54497816593886, + "grad_norm": 0.013981613330543041, + "learning_rate": 0.0006, + "loss": 4.383029937744141, + "step": 5511 + }, + { + "epoch": 76.5589519650655, + "grad_norm": 0.012930642813444138, + "learning_rate": 0.0006, + "loss": 4.296136856079102, + "step": 5512 + }, + { + "epoch": 76.57292576419214, + "grad_norm": 0.013060989789664745, + "learning_rate": 0.0006, + "loss": 4.346380233764648, + "step": 5513 + }, + { + "epoch": 76.58689956331878, + "grad_norm": 0.013209798373281956, + "learning_rate": 0.0006, + "loss": 4.316067218780518, + "step": 5514 + }, + { + "epoch": 76.60087336244541, + "grad_norm": 0.012423216365277767, + "learning_rate": 0.0006, + "loss": 4.370274543762207, + "step": 5515 + }, + { + "epoch": 76.61484716157206, + "grad_norm": 0.013205411843955517, + "learning_rate": 0.0006, + "loss": 4.429969310760498, + "step": 5516 + }, + { + "epoch": 76.62882096069869, + "grad_norm": 0.013049827888607979, + "learning_rate": 0.0006, + "loss": 4.378549098968506, + "step": 5517 + }, + { + "epoch": 76.64279475982532, + "grad_norm": 0.012075880542397499, + "learning_rate": 0.0006, + "loss": 4.348293304443359, + "step": 5518 + }, + { + "epoch": 76.65676855895197, + "grad_norm": 0.014932847581803799, + "learning_rate": 0.0006, + "loss": 4.282538890838623, + "step": 5519 + }, + { + "epoch": 76.6707423580786, + "grad_norm": 0.015615998767316341, + "learning_rate": 0.0006, + "loss": 4.274641036987305, + "step": 5520 + }, + { + "epoch": 76.68471615720524, + "grad_norm": 0.01424513477832079, + "learning_rate": 0.0006, + "loss": 4.243166923522949, + "step": 5521 + }, + { + "epoch": 76.69868995633188, + "grad_norm": 0.012775546871125698, + "learning_rate": 0.0006, + "loss": 4.335611820220947, + "step": 5522 + }, + { + "epoch": 76.71266375545852, + "grad_norm": 0.014461303129792213, + "learning_rate": 0.0006, + "loss": 4.246976375579834, + "step": 5523 + }, + { + "epoch": 76.72663755458515, + "grad_norm": 0.0145436255261302, + "learning_rate": 0.0006, + "loss": 4.275895118713379, + "step": 5524 + }, + { + "epoch": 76.74061135371178, + "grad_norm": 0.014055085368454456, + "learning_rate": 0.0006, + "loss": 4.331108093261719, + "step": 5525 + }, + { + "epoch": 76.75458515283843, + "grad_norm": 0.011933894827961922, + "learning_rate": 0.0006, + "loss": 4.293721675872803, + "step": 5526 + }, + { + "epoch": 76.76855895196506, + "grad_norm": 0.01277392078191042, + "learning_rate": 0.0006, + "loss": 4.312374114990234, + "step": 5527 + }, + { + "epoch": 76.78253275109171, + "grad_norm": 0.014227936044335365, + "learning_rate": 0.0006, + "loss": 4.238729000091553, + "step": 5528 + }, + { + "epoch": 76.79650655021834, + "grad_norm": 0.01488024927675724, + "learning_rate": 0.0006, + "loss": 4.24587345123291, + "step": 5529 + }, + { + "epoch": 76.81048034934499, + "grad_norm": 0.015341250225901604, + "learning_rate": 0.0006, + "loss": 4.257378578186035, + "step": 5530 + }, + { + "epoch": 76.82445414847162, + "grad_norm": 0.015179437585175037, + "learning_rate": 0.0006, + "loss": 4.2812910079956055, + "step": 5531 + }, + { + "epoch": 76.83842794759825, + "grad_norm": 0.012814724817872047, + "learning_rate": 0.0006, + "loss": 4.2309465408325195, + "step": 5532 + }, + { + "epoch": 76.8524017467249, + "grad_norm": 0.013810716569423676, + "learning_rate": 0.0006, + "loss": 4.325292110443115, + "step": 5533 + }, + { + "epoch": 76.86637554585153, + "grad_norm": 0.017468232661485672, + "learning_rate": 0.0006, + "loss": 4.243788719177246, + "step": 5534 + }, + { + "epoch": 76.88034934497817, + "grad_norm": 0.01739857904613018, + "learning_rate": 0.0006, + "loss": 4.345390796661377, + "step": 5535 + }, + { + "epoch": 76.8943231441048, + "grad_norm": 0.015329061076045036, + "learning_rate": 0.0006, + "loss": 4.415142059326172, + "step": 5536 + }, + { + "epoch": 76.90829694323143, + "grad_norm": 0.01598426327109337, + "learning_rate": 0.0006, + "loss": 4.299558639526367, + "step": 5537 + }, + { + "epoch": 76.92227074235808, + "grad_norm": 0.01621454581618309, + "learning_rate": 0.0006, + "loss": 4.284354209899902, + "step": 5538 + }, + { + "epoch": 76.93624454148471, + "grad_norm": 0.015956968069076538, + "learning_rate": 0.0006, + "loss": 4.293837547302246, + "step": 5539 + }, + { + "epoch": 76.95021834061136, + "grad_norm": 0.01776493526995182, + "learning_rate": 0.0006, + "loss": 4.229729175567627, + "step": 5540 + }, + { + "epoch": 76.96419213973799, + "grad_norm": 0.015317477285861969, + "learning_rate": 0.0006, + "loss": 4.3328447341918945, + "step": 5541 + }, + { + "epoch": 76.97816593886463, + "grad_norm": 0.015140251256525517, + "learning_rate": 0.0006, + "loss": 4.310840606689453, + "step": 5542 + }, + { + "epoch": 76.99213973799127, + "grad_norm": 0.015326812863349915, + "learning_rate": 0.0006, + "loss": 4.230832099914551, + "step": 5543 + }, + { + "epoch": 77.0, + "grad_norm": 0.01607763208448887, + "learning_rate": 0.0006, + "loss": 4.2236328125, + "step": 5544 + }, + { + "epoch": 77.0, + "eval_loss": 4.662221431732178, + "eval_runtime": 56.5194, + "eval_samples_per_second": 43.206, + "eval_steps_per_second": 1.362, + "step": 5544 + }, + { + "epoch": 77.01397379912663, + "grad_norm": 0.014839510433375835, + "learning_rate": 0.0006, + "loss": 4.254111289978027, + "step": 5545 + }, + { + "epoch": 77.02794759825328, + "grad_norm": 0.01519741676747799, + "learning_rate": 0.0006, + "loss": 4.1607184410095215, + "step": 5546 + }, + { + "epoch": 77.04192139737991, + "grad_norm": 0.015472603030502796, + "learning_rate": 0.0006, + "loss": 4.323320388793945, + "step": 5547 + }, + { + "epoch": 77.05589519650655, + "grad_norm": 0.017937948927283287, + "learning_rate": 0.0006, + "loss": 4.308490753173828, + "step": 5548 + }, + { + "epoch": 77.06986899563319, + "grad_norm": 0.020890841260552406, + "learning_rate": 0.0006, + "loss": 4.259101867675781, + "step": 5549 + }, + { + "epoch": 77.08384279475983, + "grad_norm": 0.01892857998609543, + "learning_rate": 0.0006, + "loss": 4.386970520019531, + "step": 5550 + }, + { + "epoch": 77.09781659388646, + "grad_norm": 0.016583630815148354, + "learning_rate": 0.0006, + "loss": 4.233604907989502, + "step": 5551 + }, + { + "epoch": 77.1117903930131, + "grad_norm": 0.016063468530774117, + "learning_rate": 0.0006, + "loss": 4.339544296264648, + "step": 5552 + }, + { + "epoch": 77.12576419213974, + "grad_norm": 0.018326779827475548, + "learning_rate": 0.0006, + "loss": 4.367204666137695, + "step": 5553 + }, + { + "epoch": 77.13973799126637, + "grad_norm": 0.021590933203697205, + "learning_rate": 0.0006, + "loss": 4.352685928344727, + "step": 5554 + }, + { + "epoch": 77.15371179039302, + "grad_norm": 0.019643642008304596, + "learning_rate": 0.0006, + "loss": 4.258535385131836, + "step": 5555 + }, + { + "epoch": 77.16768558951965, + "grad_norm": 0.018476149067282677, + "learning_rate": 0.0006, + "loss": 4.232568740844727, + "step": 5556 + }, + { + "epoch": 77.18165938864628, + "grad_norm": 0.01584405079483986, + "learning_rate": 0.0006, + "loss": 4.258371353149414, + "step": 5557 + }, + { + "epoch": 77.19563318777293, + "grad_norm": 0.01575201191008091, + "learning_rate": 0.0006, + "loss": 4.183218002319336, + "step": 5558 + }, + { + "epoch": 77.20960698689956, + "grad_norm": 0.016871824860572815, + "learning_rate": 0.0006, + "loss": 4.194563865661621, + "step": 5559 + }, + { + "epoch": 77.2235807860262, + "grad_norm": 0.02059333398938179, + "learning_rate": 0.0006, + "loss": 4.195281982421875, + "step": 5560 + }, + { + "epoch": 77.23755458515284, + "grad_norm": 0.022640453651547432, + "learning_rate": 0.0006, + "loss": 4.279615879058838, + "step": 5561 + }, + { + "epoch": 77.25152838427948, + "grad_norm": 0.020335959270596504, + "learning_rate": 0.0006, + "loss": 4.25040864944458, + "step": 5562 + }, + { + "epoch": 77.26550218340611, + "grad_norm": 0.020010793581604958, + "learning_rate": 0.0006, + "loss": 4.291116714477539, + "step": 5563 + }, + { + "epoch": 77.27947598253274, + "grad_norm": 0.02109045907855034, + "learning_rate": 0.0006, + "loss": 4.364088535308838, + "step": 5564 + }, + { + "epoch": 77.29344978165939, + "grad_norm": 0.019519446417689323, + "learning_rate": 0.0006, + "loss": 4.329166412353516, + "step": 5565 + }, + { + "epoch": 77.30742358078602, + "grad_norm": 0.01689804159104824, + "learning_rate": 0.0006, + "loss": 4.244176864624023, + "step": 5566 + }, + { + "epoch": 77.32139737991267, + "grad_norm": 0.01893681474030018, + "learning_rate": 0.0006, + "loss": 4.239068984985352, + "step": 5567 + }, + { + "epoch": 77.3353711790393, + "grad_norm": 0.015998193994164467, + "learning_rate": 0.0006, + "loss": 4.173398017883301, + "step": 5568 + }, + { + "epoch": 77.34934497816595, + "grad_norm": 0.013892280869185925, + "learning_rate": 0.0006, + "loss": 4.287654876708984, + "step": 5569 + }, + { + "epoch": 77.36331877729258, + "grad_norm": 0.017060643061995506, + "learning_rate": 0.0006, + "loss": 4.398447036743164, + "step": 5570 + }, + { + "epoch": 77.37729257641921, + "grad_norm": 0.015168849378824234, + "learning_rate": 0.0006, + "loss": 4.15665864944458, + "step": 5571 + }, + { + "epoch": 77.39126637554585, + "grad_norm": 0.014886523596942425, + "learning_rate": 0.0006, + "loss": 4.196673393249512, + "step": 5572 + }, + { + "epoch": 77.40524017467249, + "grad_norm": 0.016941796988248825, + "learning_rate": 0.0006, + "loss": 4.314562797546387, + "step": 5573 + }, + { + "epoch": 77.41921397379913, + "grad_norm": 0.015353812836110592, + "learning_rate": 0.0006, + "loss": 4.409579277038574, + "step": 5574 + }, + { + "epoch": 77.43318777292576, + "grad_norm": 0.015192852355539799, + "learning_rate": 0.0006, + "loss": 4.392045974731445, + "step": 5575 + }, + { + "epoch": 77.44716157205241, + "grad_norm": 0.014232569374144077, + "learning_rate": 0.0006, + "loss": 4.21639347076416, + "step": 5576 + }, + { + "epoch": 77.46113537117904, + "grad_norm": 0.014453226700425148, + "learning_rate": 0.0006, + "loss": 4.2802276611328125, + "step": 5577 + }, + { + "epoch": 77.47510917030567, + "grad_norm": 0.012881236150860786, + "learning_rate": 0.0006, + "loss": 4.375980854034424, + "step": 5578 + }, + { + "epoch": 77.48908296943232, + "grad_norm": 0.01583976298570633, + "learning_rate": 0.0006, + "loss": 4.264520645141602, + "step": 5579 + }, + { + "epoch": 77.50305676855895, + "grad_norm": 0.014171584509313107, + "learning_rate": 0.0006, + "loss": 4.225279808044434, + "step": 5580 + }, + { + "epoch": 77.5170305676856, + "grad_norm": 0.014715159311890602, + "learning_rate": 0.0006, + "loss": 4.369834899902344, + "step": 5581 + }, + { + "epoch": 77.53100436681223, + "grad_norm": 0.016554079949855804, + "learning_rate": 0.0006, + "loss": 4.2976179122924805, + "step": 5582 + }, + { + "epoch": 77.54497816593886, + "grad_norm": 0.015095747075974941, + "learning_rate": 0.0006, + "loss": 4.294407844543457, + "step": 5583 + }, + { + "epoch": 77.5589519650655, + "grad_norm": 0.012726361863315105, + "learning_rate": 0.0006, + "loss": 4.221401214599609, + "step": 5584 + }, + { + "epoch": 77.57292576419214, + "grad_norm": 0.014999309554696083, + "learning_rate": 0.0006, + "loss": 4.243929386138916, + "step": 5585 + }, + { + "epoch": 77.58689956331878, + "grad_norm": 0.016526037827134132, + "learning_rate": 0.0006, + "loss": 4.374139785766602, + "step": 5586 + }, + { + "epoch": 77.60087336244541, + "grad_norm": 0.01648339442908764, + "learning_rate": 0.0006, + "loss": 4.392425060272217, + "step": 5587 + }, + { + "epoch": 77.61484716157206, + "grad_norm": 0.015269213356077671, + "learning_rate": 0.0006, + "loss": 4.33986759185791, + "step": 5588 + }, + { + "epoch": 77.62882096069869, + "grad_norm": 0.01690507121384144, + "learning_rate": 0.0006, + "loss": 4.307476043701172, + "step": 5589 + }, + { + "epoch": 77.64279475982532, + "grad_norm": 0.016438685357570648, + "learning_rate": 0.0006, + "loss": 4.2406206130981445, + "step": 5590 + }, + { + "epoch": 77.65676855895197, + "grad_norm": 0.0157600287348032, + "learning_rate": 0.0006, + "loss": 4.314070224761963, + "step": 5591 + }, + { + "epoch": 77.6707423580786, + "grad_norm": 0.015127947553992271, + "learning_rate": 0.0006, + "loss": 4.22475528717041, + "step": 5592 + }, + { + "epoch": 77.68471615720524, + "grad_norm": 0.016632677987217903, + "learning_rate": 0.0006, + "loss": 4.26954984664917, + "step": 5593 + }, + { + "epoch": 77.69868995633188, + "grad_norm": 0.013150460086762905, + "learning_rate": 0.0006, + "loss": 4.369131088256836, + "step": 5594 + }, + { + "epoch": 77.71266375545852, + "grad_norm": 0.014513005502521992, + "learning_rate": 0.0006, + "loss": 4.3248796463012695, + "step": 5595 + }, + { + "epoch": 77.72663755458515, + "grad_norm": 0.014090991578996181, + "learning_rate": 0.0006, + "loss": 4.268381118774414, + "step": 5596 + }, + { + "epoch": 77.74061135371178, + "grad_norm": 0.01633097417652607, + "learning_rate": 0.0006, + "loss": 4.238439083099365, + "step": 5597 + }, + { + "epoch": 77.75458515283843, + "grad_norm": 0.01710933819413185, + "learning_rate": 0.0006, + "loss": 4.399697780609131, + "step": 5598 + }, + { + "epoch": 77.76855895196506, + "grad_norm": 0.019251767545938492, + "learning_rate": 0.0006, + "loss": 4.290450096130371, + "step": 5599 + }, + { + "epoch": 77.78253275109171, + "grad_norm": 0.01849171705543995, + "learning_rate": 0.0006, + "loss": 4.267734527587891, + "step": 5600 + }, + { + "epoch": 77.79650655021834, + "grad_norm": 0.014173115603625774, + "learning_rate": 0.0006, + "loss": 4.288799285888672, + "step": 5601 + }, + { + "epoch": 77.81048034934499, + "grad_norm": 0.014572718180716038, + "learning_rate": 0.0006, + "loss": 4.3790435791015625, + "step": 5602 + }, + { + "epoch": 77.82445414847162, + "grad_norm": 0.016375336796045303, + "learning_rate": 0.0006, + "loss": 4.231725692749023, + "step": 5603 + }, + { + "epoch": 77.83842794759825, + "grad_norm": 0.017326701432466507, + "learning_rate": 0.0006, + "loss": 4.211495399475098, + "step": 5604 + }, + { + "epoch": 77.8524017467249, + "grad_norm": 0.016399390995502472, + "learning_rate": 0.0006, + "loss": 4.386699199676514, + "step": 5605 + }, + { + "epoch": 77.86637554585153, + "grad_norm": 0.018269989639520645, + "learning_rate": 0.0006, + "loss": 4.254856109619141, + "step": 5606 + }, + { + "epoch": 77.88034934497817, + "grad_norm": 0.018443435430526733, + "learning_rate": 0.0006, + "loss": 4.391961097717285, + "step": 5607 + }, + { + "epoch": 77.8943231441048, + "grad_norm": 0.01784636080265045, + "learning_rate": 0.0006, + "loss": 4.273881912231445, + "step": 5608 + }, + { + "epoch": 77.90829694323143, + "grad_norm": 0.01782994158565998, + "learning_rate": 0.0006, + "loss": 4.26833963394165, + "step": 5609 + }, + { + "epoch": 77.92227074235808, + "grad_norm": 0.016260741278529167, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 5610 + }, + { + "epoch": 77.93624454148471, + "grad_norm": 0.015482903458178043, + "learning_rate": 0.0006, + "loss": 4.187826633453369, + "step": 5611 + }, + { + "epoch": 77.95021834061136, + "grad_norm": 0.015193257480859756, + "learning_rate": 0.0006, + "loss": 4.286060333251953, + "step": 5612 + }, + { + "epoch": 77.96419213973799, + "grad_norm": 0.015110115520656109, + "learning_rate": 0.0006, + "loss": 4.2828192710876465, + "step": 5613 + }, + { + "epoch": 77.97816593886463, + "grad_norm": 0.015973486006259918, + "learning_rate": 0.0006, + "loss": 4.317939758300781, + "step": 5614 + }, + { + "epoch": 77.99213973799127, + "grad_norm": 0.017363594844937325, + "learning_rate": 0.0006, + "loss": 4.287721633911133, + "step": 5615 + }, + { + "epoch": 78.0, + "grad_norm": 0.015277642756700516, + "learning_rate": 0.0006, + "loss": 4.356618881225586, + "step": 5616 + }, + { + "epoch": 78.0, + "eval_loss": 4.642678737640381, + "eval_runtime": 56.7335, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 5616 + }, + { + "epoch": 78.01397379912663, + "grad_norm": 0.016794051975011826, + "learning_rate": 0.0006, + "loss": 4.279930591583252, + "step": 5617 + }, + { + "epoch": 78.02794759825328, + "grad_norm": 0.01632673107087612, + "learning_rate": 0.0006, + "loss": 4.235943794250488, + "step": 5618 + }, + { + "epoch": 78.04192139737991, + "grad_norm": 0.01437876746058464, + "learning_rate": 0.0006, + "loss": 4.2542243003845215, + "step": 5619 + }, + { + "epoch": 78.05589519650655, + "grad_norm": 0.014993119053542614, + "learning_rate": 0.0006, + "loss": 4.262046813964844, + "step": 5620 + }, + { + "epoch": 78.06986899563319, + "grad_norm": 0.015721119940280914, + "learning_rate": 0.0006, + "loss": 4.173436164855957, + "step": 5621 + }, + { + "epoch": 78.08384279475983, + "grad_norm": 0.01334085687994957, + "learning_rate": 0.0006, + "loss": 4.322175979614258, + "step": 5622 + }, + { + "epoch": 78.09781659388646, + "grad_norm": 0.015561388805508614, + "learning_rate": 0.0006, + "loss": 4.242125988006592, + "step": 5623 + }, + { + "epoch": 78.1117903930131, + "grad_norm": 0.014418602921068668, + "learning_rate": 0.0006, + "loss": 4.266186714172363, + "step": 5624 + }, + { + "epoch": 78.12576419213974, + "grad_norm": 0.013739909045398235, + "learning_rate": 0.0006, + "loss": 4.231633186340332, + "step": 5625 + }, + { + "epoch": 78.13973799126637, + "grad_norm": 0.01402067206799984, + "learning_rate": 0.0006, + "loss": 4.188406944274902, + "step": 5626 + }, + { + "epoch": 78.15371179039302, + "grad_norm": 0.013639481738209724, + "learning_rate": 0.0006, + "loss": 4.1586594581604, + "step": 5627 + }, + { + "epoch": 78.16768558951965, + "grad_norm": 0.012858159840106964, + "learning_rate": 0.0006, + "loss": 4.360448360443115, + "step": 5628 + }, + { + "epoch": 78.18165938864628, + "grad_norm": 0.013089739717543125, + "learning_rate": 0.0006, + "loss": 4.344371795654297, + "step": 5629 + }, + { + "epoch": 78.19563318777293, + "grad_norm": 0.01382314320653677, + "learning_rate": 0.0006, + "loss": 4.20378303527832, + "step": 5630 + }, + { + "epoch": 78.20960698689956, + "grad_norm": 0.014206619001924992, + "learning_rate": 0.0006, + "loss": 4.239891529083252, + "step": 5631 + }, + { + "epoch": 78.2235807860262, + "grad_norm": 0.01472168043255806, + "learning_rate": 0.0006, + "loss": 4.282382488250732, + "step": 5632 + }, + { + "epoch": 78.23755458515284, + "grad_norm": 0.014237025752663612, + "learning_rate": 0.0006, + "loss": 4.241545677185059, + "step": 5633 + }, + { + "epoch": 78.25152838427948, + "grad_norm": 0.013231472112238407, + "learning_rate": 0.0006, + "loss": 4.184896469116211, + "step": 5634 + }, + { + "epoch": 78.26550218340611, + "grad_norm": 0.013538471423089504, + "learning_rate": 0.0006, + "loss": 4.325181007385254, + "step": 5635 + }, + { + "epoch": 78.27947598253274, + "grad_norm": 0.013541961088776588, + "learning_rate": 0.0006, + "loss": 4.368722915649414, + "step": 5636 + }, + { + "epoch": 78.29344978165939, + "grad_norm": 0.013446723110973835, + "learning_rate": 0.0006, + "loss": 4.301422595977783, + "step": 5637 + }, + { + "epoch": 78.30742358078602, + "grad_norm": 0.014753809198737144, + "learning_rate": 0.0006, + "loss": 4.259127616882324, + "step": 5638 + }, + { + "epoch": 78.32139737991267, + "grad_norm": 0.013256008736789227, + "learning_rate": 0.0006, + "loss": 4.284032821655273, + "step": 5639 + }, + { + "epoch": 78.3353711790393, + "grad_norm": 0.014272629283368587, + "learning_rate": 0.0006, + "loss": 4.234989166259766, + "step": 5640 + }, + { + "epoch": 78.34934497816595, + "grad_norm": 0.014756826683878899, + "learning_rate": 0.0006, + "loss": 4.143499374389648, + "step": 5641 + }, + { + "epoch": 78.36331877729258, + "grad_norm": 0.016549205407500267, + "learning_rate": 0.0006, + "loss": 4.311034202575684, + "step": 5642 + }, + { + "epoch": 78.37729257641921, + "grad_norm": 0.017735283821821213, + "learning_rate": 0.0006, + "loss": 4.246767520904541, + "step": 5643 + }, + { + "epoch": 78.39126637554585, + "grad_norm": 0.017370613291859627, + "learning_rate": 0.0006, + "loss": 4.2460832595825195, + "step": 5644 + }, + { + "epoch": 78.40524017467249, + "grad_norm": 0.014642846770584583, + "learning_rate": 0.0006, + "loss": 4.303811550140381, + "step": 5645 + }, + { + "epoch": 78.41921397379913, + "grad_norm": 0.013445726595818996, + "learning_rate": 0.0006, + "loss": 4.25921630859375, + "step": 5646 + }, + { + "epoch": 78.43318777292576, + "grad_norm": 0.014254065230488777, + "learning_rate": 0.0006, + "loss": 4.129446506500244, + "step": 5647 + }, + { + "epoch": 78.44716157205241, + "grad_norm": 0.01883019506931305, + "learning_rate": 0.0006, + "loss": 4.191716194152832, + "step": 5648 + }, + { + "epoch": 78.46113537117904, + "grad_norm": 0.0204437468200922, + "learning_rate": 0.0006, + "loss": 4.333085536956787, + "step": 5649 + }, + { + "epoch": 78.47510917030567, + "grad_norm": 0.01652318798005581, + "learning_rate": 0.0006, + "loss": 4.135462760925293, + "step": 5650 + }, + { + "epoch": 78.48908296943232, + "grad_norm": 0.012753551825881004, + "learning_rate": 0.0006, + "loss": 4.195051193237305, + "step": 5651 + }, + { + "epoch": 78.50305676855895, + "grad_norm": 0.01499118935316801, + "learning_rate": 0.0006, + "loss": 4.157092094421387, + "step": 5652 + }, + { + "epoch": 78.5170305676856, + "grad_norm": 0.015383945778012276, + "learning_rate": 0.0006, + "loss": 4.253818035125732, + "step": 5653 + }, + { + "epoch": 78.53100436681223, + "grad_norm": 0.015721332281827927, + "learning_rate": 0.0006, + "loss": 4.228775978088379, + "step": 5654 + }, + { + "epoch": 78.54497816593886, + "grad_norm": 0.014595463871955872, + "learning_rate": 0.0006, + "loss": 4.135197639465332, + "step": 5655 + }, + { + "epoch": 78.5589519650655, + "grad_norm": 0.015059644356369972, + "learning_rate": 0.0006, + "loss": 4.255614280700684, + "step": 5656 + }, + { + "epoch": 78.57292576419214, + "grad_norm": 0.013833344914019108, + "learning_rate": 0.0006, + "loss": 4.301250457763672, + "step": 5657 + }, + { + "epoch": 78.58689956331878, + "grad_norm": 0.013665376231074333, + "learning_rate": 0.0006, + "loss": 4.1734232902526855, + "step": 5658 + }, + { + "epoch": 78.60087336244541, + "grad_norm": 0.01539086177945137, + "learning_rate": 0.0006, + "loss": 4.221459865570068, + "step": 5659 + }, + { + "epoch": 78.61484716157206, + "grad_norm": 0.013313917443156242, + "learning_rate": 0.0006, + "loss": 4.255619049072266, + "step": 5660 + }, + { + "epoch": 78.62882096069869, + "grad_norm": 0.01447741687297821, + "learning_rate": 0.0006, + "loss": 4.298219680786133, + "step": 5661 + }, + { + "epoch": 78.64279475982532, + "grad_norm": 0.012796915136277676, + "learning_rate": 0.0006, + "loss": 4.2542405128479, + "step": 5662 + }, + { + "epoch": 78.65676855895197, + "grad_norm": 0.014058141969144344, + "learning_rate": 0.0006, + "loss": 4.284477233886719, + "step": 5663 + }, + { + "epoch": 78.6707423580786, + "grad_norm": 0.012705069966614246, + "learning_rate": 0.0006, + "loss": 4.2826666831970215, + "step": 5664 + }, + { + "epoch": 78.68471615720524, + "grad_norm": 0.013245890848338604, + "learning_rate": 0.0006, + "loss": 4.371612548828125, + "step": 5665 + }, + { + "epoch": 78.69868995633188, + "grad_norm": 0.014192677102982998, + "learning_rate": 0.0006, + "loss": 4.267344951629639, + "step": 5666 + }, + { + "epoch": 78.71266375545852, + "grad_norm": 0.015346777625381947, + "learning_rate": 0.0006, + "loss": 4.222797393798828, + "step": 5667 + }, + { + "epoch": 78.72663755458515, + "grad_norm": 0.014068621210753918, + "learning_rate": 0.0006, + "loss": 4.172964572906494, + "step": 5668 + }, + { + "epoch": 78.74061135371178, + "grad_norm": 0.014364945702254772, + "learning_rate": 0.0006, + "loss": 4.259319305419922, + "step": 5669 + }, + { + "epoch": 78.75458515283843, + "grad_norm": 0.017337387427687645, + "learning_rate": 0.0006, + "loss": 4.29533576965332, + "step": 5670 + }, + { + "epoch": 78.76855895196506, + "grad_norm": 0.017416715621948242, + "learning_rate": 0.0006, + "loss": 4.340364456176758, + "step": 5671 + }, + { + "epoch": 78.78253275109171, + "grad_norm": 0.015925846993923187, + "learning_rate": 0.0006, + "loss": 4.177454948425293, + "step": 5672 + }, + { + "epoch": 78.79650655021834, + "grad_norm": 0.013130443170666695, + "learning_rate": 0.0006, + "loss": 4.277454853057861, + "step": 5673 + }, + { + "epoch": 78.81048034934499, + "grad_norm": 0.014628687873482704, + "learning_rate": 0.0006, + "loss": 4.273757457733154, + "step": 5674 + }, + { + "epoch": 78.82445414847162, + "grad_norm": 0.01601659320294857, + "learning_rate": 0.0006, + "loss": 4.281774520874023, + "step": 5675 + }, + { + "epoch": 78.83842794759825, + "grad_norm": 0.015443485230207443, + "learning_rate": 0.0006, + "loss": 4.199937343597412, + "step": 5676 + }, + { + "epoch": 78.8524017467249, + "grad_norm": 0.013934246264398098, + "learning_rate": 0.0006, + "loss": 4.331369876861572, + "step": 5677 + }, + { + "epoch": 78.86637554585153, + "grad_norm": 0.01572083681821823, + "learning_rate": 0.0006, + "loss": 4.2828474044799805, + "step": 5678 + }, + { + "epoch": 78.88034934497817, + "grad_norm": 0.013444120064377785, + "learning_rate": 0.0006, + "loss": 4.183859825134277, + "step": 5679 + }, + { + "epoch": 78.8943231441048, + "grad_norm": 0.01591254211962223, + "learning_rate": 0.0006, + "loss": 4.295256614685059, + "step": 5680 + }, + { + "epoch": 78.90829694323143, + "grad_norm": 0.017556684091687202, + "learning_rate": 0.0006, + "loss": 4.132272720336914, + "step": 5681 + }, + { + "epoch": 78.92227074235808, + "grad_norm": 0.018184322863817215, + "learning_rate": 0.0006, + "loss": 4.251728057861328, + "step": 5682 + }, + { + "epoch": 78.93624454148471, + "grad_norm": 0.016207285225391388, + "learning_rate": 0.0006, + "loss": 4.205544471740723, + "step": 5683 + }, + { + "epoch": 78.95021834061136, + "grad_norm": 0.014478376135230064, + "learning_rate": 0.0006, + "loss": 4.182462215423584, + "step": 5684 + }, + { + "epoch": 78.96419213973799, + "grad_norm": 0.014255235902965069, + "learning_rate": 0.0006, + "loss": 4.253366947174072, + "step": 5685 + }, + { + "epoch": 78.97816593886463, + "grad_norm": 0.01524396613240242, + "learning_rate": 0.0006, + "loss": 4.221786022186279, + "step": 5686 + }, + { + "epoch": 78.99213973799127, + "grad_norm": 0.015057693235576153, + "learning_rate": 0.0006, + "loss": 4.208454608917236, + "step": 5687 + }, + { + "epoch": 79.0, + "grad_norm": 0.015478022396564484, + "learning_rate": 0.0006, + "loss": 4.336383819580078, + "step": 5688 + }, + { + "epoch": 79.0, + "eval_loss": 4.641749382019043, + "eval_runtime": 56.8167, + "eval_samples_per_second": 42.98, + "eval_steps_per_second": 1.355, + "step": 5688 + }, + { + "epoch": 79.01397379912663, + "grad_norm": 0.014143316075205803, + "learning_rate": 0.0006, + "loss": 4.260974884033203, + "step": 5689 + }, + { + "epoch": 79.02794759825328, + "grad_norm": 0.013120883144438267, + "learning_rate": 0.0006, + "loss": 4.139886856079102, + "step": 5690 + }, + { + "epoch": 79.04192139737991, + "grad_norm": 0.015019661746919155, + "learning_rate": 0.0006, + "loss": 4.267004013061523, + "step": 5691 + }, + { + "epoch": 79.05589519650655, + "grad_norm": 0.014891284517943859, + "learning_rate": 0.0006, + "loss": 4.205019950866699, + "step": 5692 + }, + { + "epoch": 79.06986899563319, + "grad_norm": 0.015527728945016861, + "learning_rate": 0.0006, + "loss": 4.1513566970825195, + "step": 5693 + }, + { + "epoch": 79.08384279475983, + "grad_norm": 0.017527181655168533, + "learning_rate": 0.0006, + "loss": 4.148828983306885, + "step": 5694 + }, + { + "epoch": 79.09781659388646, + "grad_norm": 0.01632206328213215, + "learning_rate": 0.0006, + "loss": 4.195789813995361, + "step": 5695 + }, + { + "epoch": 79.1117903930131, + "grad_norm": 0.01590503193438053, + "learning_rate": 0.0006, + "loss": 4.199155330657959, + "step": 5696 + }, + { + "epoch": 79.12576419213974, + "grad_norm": 0.016489997506141663, + "learning_rate": 0.0006, + "loss": 4.155154705047607, + "step": 5697 + }, + { + "epoch": 79.13973799126637, + "grad_norm": 0.0161368977278471, + "learning_rate": 0.0006, + "loss": 4.164384841918945, + "step": 5698 + }, + { + "epoch": 79.15371179039302, + "grad_norm": 0.018574809655547142, + "learning_rate": 0.0006, + "loss": 4.1898193359375, + "step": 5699 + }, + { + "epoch": 79.16768558951965, + "grad_norm": 0.018987983465194702, + "learning_rate": 0.0006, + "loss": 4.219719409942627, + "step": 5700 + }, + { + "epoch": 79.18165938864628, + "grad_norm": 0.01668606884777546, + "learning_rate": 0.0006, + "loss": 4.152734756469727, + "step": 5701 + }, + { + "epoch": 79.19563318777293, + "grad_norm": 0.014046299271285534, + "learning_rate": 0.0006, + "loss": 4.152964115142822, + "step": 5702 + }, + { + "epoch": 79.20960698689956, + "grad_norm": 0.01628032699227333, + "learning_rate": 0.0006, + "loss": 4.299479961395264, + "step": 5703 + }, + { + "epoch": 79.2235807860262, + "grad_norm": 0.017998795956373215, + "learning_rate": 0.0006, + "loss": 4.231203079223633, + "step": 5704 + }, + { + "epoch": 79.23755458515284, + "grad_norm": 0.014799647964537144, + "learning_rate": 0.0006, + "loss": 4.3530192375183105, + "step": 5705 + }, + { + "epoch": 79.25152838427948, + "grad_norm": 0.014640678651630878, + "learning_rate": 0.0006, + "loss": 4.287322998046875, + "step": 5706 + }, + { + "epoch": 79.26550218340611, + "grad_norm": 0.0172616895288229, + "learning_rate": 0.0006, + "loss": 4.148009777069092, + "step": 5707 + }, + { + "epoch": 79.27947598253274, + "grad_norm": 0.01811790093779564, + "learning_rate": 0.0006, + "loss": 4.269433975219727, + "step": 5708 + }, + { + "epoch": 79.29344978165939, + "grad_norm": 0.01764000579714775, + "learning_rate": 0.0006, + "loss": 4.292638778686523, + "step": 5709 + }, + { + "epoch": 79.30742358078602, + "grad_norm": 0.018030935898423195, + "learning_rate": 0.0006, + "loss": 4.267889976501465, + "step": 5710 + }, + { + "epoch": 79.32139737991267, + "grad_norm": 0.017860891297459602, + "learning_rate": 0.0006, + "loss": 4.211110591888428, + "step": 5711 + }, + { + "epoch": 79.3353711790393, + "grad_norm": 0.017295073717832565, + "learning_rate": 0.0006, + "loss": 4.224645614624023, + "step": 5712 + }, + { + "epoch": 79.34934497816595, + "grad_norm": 0.015891101211309433, + "learning_rate": 0.0006, + "loss": 4.224797248840332, + "step": 5713 + }, + { + "epoch": 79.36331877729258, + "grad_norm": 0.014680417254567146, + "learning_rate": 0.0006, + "loss": 4.182242393493652, + "step": 5714 + }, + { + "epoch": 79.37729257641921, + "grad_norm": 0.016621140763163567, + "learning_rate": 0.0006, + "loss": 4.186334609985352, + "step": 5715 + }, + { + "epoch": 79.39126637554585, + "grad_norm": 0.016207249835133553, + "learning_rate": 0.0006, + "loss": 4.277074813842773, + "step": 5716 + }, + { + "epoch": 79.40524017467249, + "grad_norm": 0.015375478193163872, + "learning_rate": 0.0006, + "loss": 4.154317855834961, + "step": 5717 + }, + { + "epoch": 79.41921397379913, + "grad_norm": 0.013453228399157524, + "learning_rate": 0.0006, + "loss": 4.265270233154297, + "step": 5718 + }, + { + "epoch": 79.43318777292576, + "grad_norm": 0.014312694780528545, + "learning_rate": 0.0006, + "loss": 4.232220649719238, + "step": 5719 + }, + { + "epoch": 79.44716157205241, + "grad_norm": 0.014759554527699947, + "learning_rate": 0.0006, + "loss": 4.272275924682617, + "step": 5720 + }, + { + "epoch": 79.46113537117904, + "grad_norm": 0.015885505825281143, + "learning_rate": 0.0006, + "loss": 4.2229719161987305, + "step": 5721 + }, + { + "epoch": 79.47510917030567, + "grad_norm": 0.016516581177711487, + "learning_rate": 0.0006, + "loss": 4.200624465942383, + "step": 5722 + }, + { + "epoch": 79.48908296943232, + "grad_norm": 0.018254555761814117, + "learning_rate": 0.0006, + "loss": 4.19648551940918, + "step": 5723 + }, + { + "epoch": 79.50305676855895, + "grad_norm": 0.01467090006917715, + "learning_rate": 0.0006, + "loss": 4.135190486907959, + "step": 5724 + }, + { + "epoch": 79.5170305676856, + "grad_norm": 0.013474797829985619, + "learning_rate": 0.0006, + "loss": 4.311361312866211, + "step": 5725 + }, + { + "epoch": 79.53100436681223, + "grad_norm": 0.01742764748632908, + "learning_rate": 0.0006, + "loss": 4.1845903396606445, + "step": 5726 + }, + { + "epoch": 79.54497816593886, + "grad_norm": 0.018786882981657982, + "learning_rate": 0.0006, + "loss": 4.163942337036133, + "step": 5727 + }, + { + "epoch": 79.5589519650655, + "grad_norm": 0.017831219360232353, + "learning_rate": 0.0006, + "loss": 4.236685276031494, + "step": 5728 + }, + { + "epoch": 79.57292576419214, + "grad_norm": 0.016334986314177513, + "learning_rate": 0.0006, + "loss": 4.255853652954102, + "step": 5729 + }, + { + "epoch": 79.58689956331878, + "grad_norm": 0.0138487434014678, + "learning_rate": 0.0006, + "loss": 4.227874279022217, + "step": 5730 + }, + { + "epoch": 79.60087336244541, + "grad_norm": 0.01531760673969984, + "learning_rate": 0.0006, + "loss": 4.221432685852051, + "step": 5731 + }, + { + "epoch": 79.61484716157206, + "grad_norm": 0.01603361964225769, + "learning_rate": 0.0006, + "loss": 4.278326511383057, + "step": 5732 + }, + { + "epoch": 79.62882096069869, + "grad_norm": 0.013149895705282688, + "learning_rate": 0.0006, + "loss": 4.2474517822265625, + "step": 5733 + }, + { + "epoch": 79.64279475982532, + "grad_norm": 0.013428786769509315, + "learning_rate": 0.0006, + "loss": 4.1701178550720215, + "step": 5734 + }, + { + "epoch": 79.65676855895197, + "grad_norm": 0.014305585995316505, + "learning_rate": 0.0006, + "loss": 4.250778675079346, + "step": 5735 + }, + { + "epoch": 79.6707423580786, + "grad_norm": 0.014544566161930561, + "learning_rate": 0.0006, + "loss": 4.243236541748047, + "step": 5736 + }, + { + "epoch": 79.68471615720524, + "grad_norm": 0.013524414040148258, + "learning_rate": 0.0006, + "loss": 4.255457878112793, + "step": 5737 + }, + { + "epoch": 79.69868995633188, + "grad_norm": 0.014750641770660877, + "learning_rate": 0.0006, + "loss": 4.202816486358643, + "step": 5738 + }, + { + "epoch": 79.71266375545852, + "grad_norm": 0.015090773813426495, + "learning_rate": 0.0006, + "loss": 4.3053178787231445, + "step": 5739 + }, + { + "epoch": 79.72663755458515, + "grad_norm": 0.01606169529259205, + "learning_rate": 0.0006, + "loss": 4.224017143249512, + "step": 5740 + }, + { + "epoch": 79.74061135371178, + "grad_norm": 0.016303520649671555, + "learning_rate": 0.0006, + "loss": 4.190760612487793, + "step": 5741 + }, + { + "epoch": 79.75458515283843, + "grad_norm": 0.01465672068297863, + "learning_rate": 0.0006, + "loss": 4.24506950378418, + "step": 5742 + }, + { + "epoch": 79.76855895196506, + "grad_norm": 0.012616301886737347, + "learning_rate": 0.0006, + "loss": 4.174039840698242, + "step": 5743 + }, + { + "epoch": 79.78253275109171, + "grad_norm": 0.01358284242451191, + "learning_rate": 0.0006, + "loss": 4.236380577087402, + "step": 5744 + }, + { + "epoch": 79.79650655021834, + "grad_norm": 0.01557177770882845, + "learning_rate": 0.0006, + "loss": 4.102962493896484, + "step": 5745 + }, + { + "epoch": 79.81048034934499, + "grad_norm": 0.015584317035973072, + "learning_rate": 0.0006, + "loss": 4.242603302001953, + "step": 5746 + }, + { + "epoch": 79.82445414847162, + "grad_norm": 0.012597334571182728, + "learning_rate": 0.0006, + "loss": 4.202791690826416, + "step": 5747 + }, + { + "epoch": 79.83842794759825, + "grad_norm": 0.013722898438572884, + "learning_rate": 0.0006, + "loss": 4.204328536987305, + "step": 5748 + }, + { + "epoch": 79.8524017467249, + "grad_norm": 0.014669225551187992, + "learning_rate": 0.0006, + "loss": 4.164331912994385, + "step": 5749 + }, + { + "epoch": 79.86637554585153, + "grad_norm": 0.015012740157544613, + "learning_rate": 0.0006, + "loss": 4.260565280914307, + "step": 5750 + }, + { + "epoch": 79.88034934497817, + "grad_norm": 0.016026340425014496, + "learning_rate": 0.0006, + "loss": 4.299723148345947, + "step": 5751 + }, + { + "epoch": 79.8943231441048, + "grad_norm": 0.01626049168407917, + "learning_rate": 0.0006, + "loss": 4.207252502441406, + "step": 5752 + }, + { + "epoch": 79.90829694323143, + "grad_norm": 0.015415354631841183, + "learning_rate": 0.0006, + "loss": 4.151934623718262, + "step": 5753 + }, + { + "epoch": 79.92227074235808, + "grad_norm": 0.014158242382109165, + "learning_rate": 0.0006, + "loss": 4.3461995124816895, + "step": 5754 + }, + { + "epoch": 79.93624454148471, + "grad_norm": 0.01585628278553486, + "learning_rate": 0.0006, + "loss": 4.334761619567871, + "step": 5755 + }, + { + "epoch": 79.95021834061136, + "grad_norm": 0.013710759580135345, + "learning_rate": 0.0006, + "loss": 4.180401802062988, + "step": 5756 + }, + { + "epoch": 79.96419213973799, + "grad_norm": 0.014159204438328743, + "learning_rate": 0.0006, + "loss": 4.322331428527832, + "step": 5757 + }, + { + "epoch": 79.97816593886463, + "grad_norm": 0.015153195708990097, + "learning_rate": 0.0006, + "loss": 4.287478923797607, + "step": 5758 + }, + { + "epoch": 79.99213973799127, + "grad_norm": 0.014100227504968643, + "learning_rate": 0.0006, + "loss": 4.231807708740234, + "step": 5759 + }, + { + "epoch": 80.0, + "grad_norm": 0.015248682349920273, + "learning_rate": 0.0006, + "loss": 4.261106014251709, + "step": 5760 + }, + { + "epoch": 80.0, + "eval_loss": 4.610264301300049, + "eval_runtime": 56.514, + "eval_samples_per_second": 43.211, + "eval_steps_per_second": 1.362, + "step": 5760 + }, + { + "epoch": 80.01397379912663, + "grad_norm": 0.014730271883308887, + "learning_rate": 0.0006, + "loss": 4.108556747436523, + "step": 5761 + }, + { + "epoch": 80.02794759825328, + "grad_norm": 0.014575716108083725, + "learning_rate": 0.0006, + "loss": 4.169658660888672, + "step": 5762 + }, + { + "epoch": 80.04192139737991, + "grad_norm": 0.014706958085298538, + "learning_rate": 0.0006, + "loss": 4.252662658691406, + "step": 5763 + }, + { + "epoch": 80.05589519650655, + "grad_norm": 0.01561224739998579, + "learning_rate": 0.0006, + "loss": 4.193892955780029, + "step": 5764 + }, + { + "epoch": 80.06986899563319, + "grad_norm": 0.014702324755489826, + "learning_rate": 0.0006, + "loss": 4.1889238357543945, + "step": 5765 + }, + { + "epoch": 80.08384279475983, + "grad_norm": 0.015565741807222366, + "learning_rate": 0.0006, + "loss": 4.243549823760986, + "step": 5766 + }, + { + "epoch": 80.09781659388646, + "grad_norm": 0.01647181063890457, + "learning_rate": 0.0006, + "loss": 4.275446891784668, + "step": 5767 + }, + { + "epoch": 80.1117903930131, + "grad_norm": 0.018177716061472893, + "learning_rate": 0.0006, + "loss": 4.12040376663208, + "step": 5768 + }, + { + "epoch": 80.12576419213974, + "grad_norm": 0.01745191030204296, + "learning_rate": 0.0006, + "loss": 4.140371322631836, + "step": 5769 + }, + { + "epoch": 80.13973799126637, + "grad_norm": 0.01568993367254734, + "learning_rate": 0.0006, + "loss": 4.258280277252197, + "step": 5770 + }, + { + "epoch": 80.15371179039302, + "grad_norm": 0.01636597141623497, + "learning_rate": 0.0006, + "loss": 4.179052352905273, + "step": 5771 + }, + { + "epoch": 80.16768558951965, + "grad_norm": 0.016839100047945976, + "learning_rate": 0.0006, + "loss": 4.127636432647705, + "step": 5772 + }, + { + "epoch": 80.18165938864628, + "grad_norm": 0.015471016988158226, + "learning_rate": 0.0006, + "loss": 4.181325912475586, + "step": 5773 + }, + { + "epoch": 80.19563318777293, + "grad_norm": 0.015836089849472046, + "learning_rate": 0.0006, + "loss": 4.142168998718262, + "step": 5774 + }, + { + "epoch": 80.20960698689956, + "grad_norm": 0.014782954007387161, + "learning_rate": 0.0006, + "loss": 4.308799743652344, + "step": 5775 + }, + { + "epoch": 80.2235807860262, + "grad_norm": 0.014040893875062466, + "learning_rate": 0.0006, + "loss": 4.216936111450195, + "step": 5776 + }, + { + "epoch": 80.23755458515284, + "grad_norm": 0.01420280709862709, + "learning_rate": 0.0006, + "loss": 4.235230445861816, + "step": 5777 + }, + { + "epoch": 80.25152838427948, + "grad_norm": 0.013341007754206657, + "learning_rate": 0.0006, + "loss": 4.369109153747559, + "step": 5778 + }, + { + "epoch": 80.26550218340611, + "grad_norm": 0.014208327978849411, + "learning_rate": 0.0006, + "loss": 4.219112396240234, + "step": 5779 + }, + { + "epoch": 80.27947598253274, + "grad_norm": 0.013825052417814732, + "learning_rate": 0.0006, + "loss": 4.1983489990234375, + "step": 5780 + }, + { + "epoch": 80.29344978165939, + "grad_norm": 0.012916515581309795, + "learning_rate": 0.0006, + "loss": 4.204582691192627, + "step": 5781 + }, + { + "epoch": 80.30742358078602, + "grad_norm": 0.013071243651211262, + "learning_rate": 0.0006, + "loss": 4.17643928527832, + "step": 5782 + }, + { + "epoch": 80.32139737991267, + "grad_norm": 0.015507174655795097, + "learning_rate": 0.0006, + "loss": 4.172976016998291, + "step": 5783 + }, + { + "epoch": 80.3353711790393, + "grad_norm": 0.014967243187129498, + "learning_rate": 0.0006, + "loss": 4.141168594360352, + "step": 5784 + }, + { + "epoch": 80.34934497816595, + "grad_norm": 0.015629924833774567, + "learning_rate": 0.0006, + "loss": 4.253526210784912, + "step": 5785 + }, + { + "epoch": 80.36331877729258, + "grad_norm": 0.01607954502105713, + "learning_rate": 0.0006, + "loss": 4.249395370483398, + "step": 5786 + }, + { + "epoch": 80.37729257641921, + "grad_norm": 0.01500706560909748, + "learning_rate": 0.0006, + "loss": 4.279012680053711, + "step": 5787 + }, + { + "epoch": 80.39126637554585, + "grad_norm": 0.014343900606036186, + "learning_rate": 0.0006, + "loss": 4.268509864807129, + "step": 5788 + }, + { + "epoch": 80.40524017467249, + "grad_norm": 0.014682197012007236, + "learning_rate": 0.0006, + "loss": 4.206334114074707, + "step": 5789 + }, + { + "epoch": 80.41921397379913, + "grad_norm": 0.014989163726568222, + "learning_rate": 0.0006, + "loss": 4.143601417541504, + "step": 5790 + }, + { + "epoch": 80.43318777292576, + "grad_norm": 0.015001608058810234, + "learning_rate": 0.0006, + "loss": 4.164111137390137, + "step": 5791 + }, + { + "epoch": 80.44716157205241, + "grad_norm": 0.015222722664475441, + "learning_rate": 0.0006, + "loss": 4.2369914054870605, + "step": 5792 + }, + { + "epoch": 80.46113537117904, + "grad_norm": 0.014126183465123177, + "learning_rate": 0.0006, + "loss": 4.2209930419921875, + "step": 5793 + }, + { + "epoch": 80.47510917030567, + "grad_norm": 0.013065959326922894, + "learning_rate": 0.0006, + "loss": 4.231184959411621, + "step": 5794 + }, + { + "epoch": 80.48908296943232, + "grad_norm": 0.013833962380886078, + "learning_rate": 0.0006, + "loss": 4.226434707641602, + "step": 5795 + }, + { + "epoch": 80.50305676855895, + "grad_norm": 0.01387125626206398, + "learning_rate": 0.0006, + "loss": 4.241328716278076, + "step": 5796 + }, + { + "epoch": 80.5170305676856, + "grad_norm": 0.014145063236355782, + "learning_rate": 0.0006, + "loss": 4.2268805503845215, + "step": 5797 + }, + { + "epoch": 80.53100436681223, + "grad_norm": 0.012990240007638931, + "learning_rate": 0.0006, + "loss": 4.1718339920043945, + "step": 5798 + }, + { + "epoch": 80.54497816593886, + "grad_norm": 0.014185411855578423, + "learning_rate": 0.0006, + "loss": 4.232397556304932, + "step": 5799 + }, + { + "epoch": 80.5589519650655, + "grad_norm": 0.014260523952543736, + "learning_rate": 0.0006, + "loss": 4.282734394073486, + "step": 5800 + }, + { + "epoch": 80.57292576419214, + "grad_norm": 0.012542120181024075, + "learning_rate": 0.0006, + "loss": 4.134551525115967, + "step": 5801 + }, + { + "epoch": 80.58689956331878, + "grad_norm": 0.013078657910227776, + "learning_rate": 0.0006, + "loss": 4.229815483093262, + "step": 5802 + }, + { + "epoch": 80.60087336244541, + "grad_norm": 0.012821056880056858, + "learning_rate": 0.0006, + "loss": 4.208278179168701, + "step": 5803 + }, + { + "epoch": 80.61484716157206, + "grad_norm": 0.012964180670678616, + "learning_rate": 0.0006, + "loss": 4.188477039337158, + "step": 5804 + }, + { + "epoch": 80.62882096069869, + "grad_norm": 0.01631746254861355, + "learning_rate": 0.0006, + "loss": 4.3214497566223145, + "step": 5805 + }, + { + "epoch": 80.64279475982532, + "grad_norm": 0.01726640947163105, + "learning_rate": 0.0006, + "loss": 4.279359817504883, + "step": 5806 + }, + { + "epoch": 80.65676855895197, + "grad_norm": 0.018778258934617043, + "learning_rate": 0.0006, + "loss": 4.294047832489014, + "step": 5807 + }, + { + "epoch": 80.6707423580786, + "grad_norm": 0.01796831749379635, + "learning_rate": 0.0006, + "loss": 4.314162254333496, + "step": 5808 + }, + { + "epoch": 80.68471615720524, + "grad_norm": 0.016495538875460625, + "learning_rate": 0.0006, + "loss": 4.273871421813965, + "step": 5809 + }, + { + "epoch": 80.69868995633188, + "grad_norm": 0.013590306043624878, + "learning_rate": 0.0006, + "loss": 4.14961051940918, + "step": 5810 + }, + { + "epoch": 80.71266375545852, + "grad_norm": 0.014670426957309246, + "learning_rate": 0.0006, + "loss": 4.211853981018066, + "step": 5811 + }, + { + "epoch": 80.72663755458515, + "grad_norm": 0.01641341857612133, + "learning_rate": 0.0006, + "loss": 4.211948871612549, + "step": 5812 + }, + { + "epoch": 80.74061135371178, + "grad_norm": 0.016365470364689827, + "learning_rate": 0.0006, + "loss": 4.200774192810059, + "step": 5813 + }, + { + "epoch": 80.75458515283843, + "grad_norm": 0.01613970287144184, + "learning_rate": 0.0006, + "loss": 4.288322448730469, + "step": 5814 + }, + { + "epoch": 80.76855895196506, + "grad_norm": 0.016742777079343796, + "learning_rate": 0.0006, + "loss": 4.23801326751709, + "step": 5815 + }, + { + "epoch": 80.78253275109171, + "grad_norm": 0.015351069159805775, + "learning_rate": 0.0006, + "loss": 4.263832092285156, + "step": 5816 + }, + { + "epoch": 80.79650655021834, + "grad_norm": 0.013435336761176586, + "learning_rate": 0.0006, + "loss": 4.160857200622559, + "step": 5817 + }, + { + "epoch": 80.81048034934499, + "grad_norm": 0.01474886853247881, + "learning_rate": 0.0006, + "loss": 4.216025352478027, + "step": 5818 + }, + { + "epoch": 80.82445414847162, + "grad_norm": 0.01586935855448246, + "learning_rate": 0.0006, + "loss": 4.256402969360352, + "step": 5819 + }, + { + "epoch": 80.83842794759825, + "grad_norm": 0.014596572145819664, + "learning_rate": 0.0006, + "loss": 4.142584800720215, + "step": 5820 + }, + { + "epoch": 80.8524017467249, + "grad_norm": 0.01501593366265297, + "learning_rate": 0.0006, + "loss": 4.250462055206299, + "step": 5821 + }, + { + "epoch": 80.86637554585153, + "grad_norm": 0.014773511327803135, + "learning_rate": 0.0006, + "loss": 4.253305912017822, + "step": 5822 + }, + { + "epoch": 80.88034934497817, + "grad_norm": 0.013246326707303524, + "learning_rate": 0.0006, + "loss": 4.170881271362305, + "step": 5823 + }, + { + "epoch": 80.8943231441048, + "grad_norm": 0.013013756833970547, + "learning_rate": 0.0006, + "loss": 4.264721870422363, + "step": 5824 + }, + { + "epoch": 80.90829694323143, + "grad_norm": 0.013155903667211533, + "learning_rate": 0.0006, + "loss": 4.274667739868164, + "step": 5825 + }, + { + "epoch": 80.92227074235808, + "grad_norm": 0.01448952779173851, + "learning_rate": 0.0006, + "loss": 4.24777889251709, + "step": 5826 + }, + { + "epoch": 80.93624454148471, + "grad_norm": 0.01415231078863144, + "learning_rate": 0.0006, + "loss": 4.22014856338501, + "step": 5827 + }, + { + "epoch": 80.95021834061136, + "grad_norm": 0.014796273782849312, + "learning_rate": 0.0006, + "loss": 4.22036600112915, + "step": 5828 + }, + { + "epoch": 80.96419213973799, + "grad_norm": 0.013005326502025127, + "learning_rate": 0.0006, + "loss": 4.276123046875, + "step": 5829 + }, + { + "epoch": 80.97816593886463, + "grad_norm": 0.012851177714765072, + "learning_rate": 0.0006, + "loss": 4.248575687408447, + "step": 5830 + }, + { + "epoch": 80.99213973799127, + "grad_norm": 0.013740002177655697, + "learning_rate": 0.0006, + "loss": 4.051140308380127, + "step": 5831 + }, + { + "epoch": 81.0, + "grad_norm": 0.015296104364097118, + "learning_rate": 0.0006, + "loss": 4.251155853271484, + "step": 5832 + }, + { + "epoch": 81.0, + "eval_loss": 4.594127655029297, + "eval_runtime": 56.0959, + "eval_samples_per_second": 43.533, + "eval_steps_per_second": 1.373, + "step": 5832 + }, + { + "epoch": 81.01397379912663, + "grad_norm": 0.015804558992385864, + "learning_rate": 0.0006, + "loss": 4.152894020080566, + "step": 5833 + }, + { + "epoch": 81.02794759825328, + "grad_norm": 0.017000854015350342, + "learning_rate": 0.0006, + "loss": 4.266067981719971, + "step": 5834 + }, + { + "epoch": 81.04192139737991, + "grad_norm": 0.016351008787751198, + "learning_rate": 0.0006, + "loss": 4.192703723907471, + "step": 5835 + }, + { + "epoch": 81.05589519650655, + "grad_norm": 0.014703895896673203, + "learning_rate": 0.0006, + "loss": 4.23554801940918, + "step": 5836 + }, + { + "epoch": 81.06986899563319, + "grad_norm": 0.01619596965610981, + "learning_rate": 0.0006, + "loss": 4.233000755310059, + "step": 5837 + }, + { + "epoch": 81.08384279475983, + "grad_norm": 0.01815493032336235, + "learning_rate": 0.0006, + "loss": 4.200490951538086, + "step": 5838 + }, + { + "epoch": 81.09781659388646, + "grad_norm": 0.018305106088519096, + "learning_rate": 0.0006, + "loss": 4.154242515563965, + "step": 5839 + }, + { + "epoch": 81.1117903930131, + "grad_norm": 0.0200856514275074, + "learning_rate": 0.0006, + "loss": 4.25820779800415, + "step": 5840 + }, + { + "epoch": 81.12576419213974, + "grad_norm": 0.02101440541446209, + "learning_rate": 0.0006, + "loss": 4.168209075927734, + "step": 5841 + }, + { + "epoch": 81.13973799126637, + "grad_norm": 0.022829681634902954, + "learning_rate": 0.0006, + "loss": 4.25560998916626, + "step": 5842 + }, + { + "epoch": 81.15371179039302, + "grad_norm": 0.0195082426071167, + "learning_rate": 0.0006, + "loss": 4.2035627365112305, + "step": 5843 + }, + { + "epoch": 81.16768558951965, + "grad_norm": 0.01716248318552971, + "learning_rate": 0.0006, + "loss": 4.11322021484375, + "step": 5844 + }, + { + "epoch": 81.18165938864628, + "grad_norm": 0.02034664712846279, + "learning_rate": 0.0006, + "loss": 4.179389953613281, + "step": 5845 + }, + { + "epoch": 81.19563318777293, + "grad_norm": 0.023578710854053497, + "learning_rate": 0.0006, + "loss": 4.265695571899414, + "step": 5846 + }, + { + "epoch": 81.20960698689956, + "grad_norm": 0.019244182854890823, + "learning_rate": 0.0006, + "loss": 4.292043209075928, + "step": 5847 + }, + { + "epoch": 81.2235807860262, + "grad_norm": 0.01577291637659073, + "learning_rate": 0.0006, + "loss": 4.250776290893555, + "step": 5848 + }, + { + "epoch": 81.23755458515284, + "grad_norm": 0.01617090217769146, + "learning_rate": 0.0006, + "loss": 4.193098068237305, + "step": 5849 + }, + { + "epoch": 81.25152838427948, + "grad_norm": 0.016090184450149536, + "learning_rate": 0.0006, + "loss": 4.112357139587402, + "step": 5850 + }, + { + "epoch": 81.26550218340611, + "grad_norm": 0.01441365759819746, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 5851 + }, + { + "epoch": 81.27947598253274, + "grad_norm": 0.014122162945568562, + "learning_rate": 0.0006, + "loss": 4.276175022125244, + "step": 5852 + }, + { + "epoch": 81.29344978165939, + "grad_norm": 0.014859487302601337, + "learning_rate": 0.0006, + "loss": 4.094699859619141, + "step": 5853 + }, + { + "epoch": 81.30742358078602, + "grad_norm": 0.013624267652630806, + "learning_rate": 0.0006, + "loss": 4.2648773193359375, + "step": 5854 + }, + { + "epoch": 81.32139737991267, + "grad_norm": 0.01417913381010294, + "learning_rate": 0.0006, + "loss": 4.228567123413086, + "step": 5855 + }, + { + "epoch": 81.3353711790393, + "grad_norm": 0.014560805633664131, + "learning_rate": 0.0006, + "loss": 4.148930549621582, + "step": 5856 + }, + { + "epoch": 81.34934497816595, + "grad_norm": 0.013662154786288738, + "learning_rate": 0.0006, + "loss": 4.192107677459717, + "step": 5857 + }, + { + "epoch": 81.36331877729258, + "grad_norm": 0.013434980995953083, + "learning_rate": 0.0006, + "loss": 4.162052154541016, + "step": 5858 + }, + { + "epoch": 81.37729257641921, + "grad_norm": 0.01300742756575346, + "learning_rate": 0.0006, + "loss": 4.262385368347168, + "step": 5859 + }, + { + "epoch": 81.39126637554585, + "grad_norm": 0.012544351629912853, + "learning_rate": 0.0006, + "loss": 4.215371131896973, + "step": 5860 + }, + { + "epoch": 81.40524017467249, + "grad_norm": 0.013663645833730698, + "learning_rate": 0.0006, + "loss": 4.247312545776367, + "step": 5861 + }, + { + "epoch": 81.41921397379913, + "grad_norm": 0.013127296231687069, + "learning_rate": 0.0006, + "loss": 4.087685585021973, + "step": 5862 + }, + { + "epoch": 81.43318777292576, + "grad_norm": 0.014081938192248344, + "learning_rate": 0.0006, + "loss": 4.272648811340332, + "step": 5863 + }, + { + "epoch": 81.44716157205241, + "grad_norm": 0.015470641665160656, + "learning_rate": 0.0006, + "loss": 4.278578758239746, + "step": 5864 + }, + { + "epoch": 81.46113537117904, + "grad_norm": 0.016524532809853554, + "learning_rate": 0.0006, + "loss": 4.118268013000488, + "step": 5865 + }, + { + "epoch": 81.47510917030567, + "grad_norm": 0.016242722049355507, + "learning_rate": 0.0006, + "loss": 4.182405471801758, + "step": 5866 + }, + { + "epoch": 81.48908296943232, + "grad_norm": 0.01557017583400011, + "learning_rate": 0.0006, + "loss": 4.229862213134766, + "step": 5867 + }, + { + "epoch": 81.50305676855895, + "grad_norm": 0.01562870666384697, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 5868 + }, + { + "epoch": 81.5170305676856, + "grad_norm": 0.018606074154376984, + "learning_rate": 0.0006, + "loss": 4.148681640625, + "step": 5869 + }, + { + "epoch": 81.53100436681223, + "grad_norm": 0.017740730196237564, + "learning_rate": 0.0006, + "loss": 4.321932792663574, + "step": 5870 + }, + { + "epoch": 81.54497816593886, + "grad_norm": 0.014517677947878838, + "learning_rate": 0.0006, + "loss": 4.22862434387207, + "step": 5871 + }, + { + "epoch": 81.5589519650655, + "grad_norm": 0.01389714702963829, + "learning_rate": 0.0006, + "loss": 4.198000907897949, + "step": 5872 + }, + { + "epoch": 81.57292576419214, + "grad_norm": 0.015348682180047035, + "learning_rate": 0.0006, + "loss": 4.293485164642334, + "step": 5873 + }, + { + "epoch": 81.58689956331878, + "grad_norm": 0.015965791419148445, + "learning_rate": 0.0006, + "loss": 4.23481559753418, + "step": 5874 + }, + { + "epoch": 81.60087336244541, + "grad_norm": 0.015662726014852524, + "learning_rate": 0.0006, + "loss": 4.158148765563965, + "step": 5875 + }, + { + "epoch": 81.61484716157206, + "grad_norm": 0.014402080327272415, + "learning_rate": 0.0006, + "loss": 4.132723808288574, + "step": 5876 + }, + { + "epoch": 81.62882096069869, + "grad_norm": 0.016846125945448875, + "learning_rate": 0.0006, + "loss": 4.1654462814331055, + "step": 5877 + }, + { + "epoch": 81.64279475982532, + "grad_norm": 0.0196097232401371, + "learning_rate": 0.0006, + "loss": 4.228450298309326, + "step": 5878 + }, + { + "epoch": 81.65676855895197, + "grad_norm": 0.016005294397473335, + "learning_rate": 0.0006, + "loss": 4.230222225189209, + "step": 5879 + }, + { + "epoch": 81.6707423580786, + "grad_norm": 0.014685059897601604, + "learning_rate": 0.0006, + "loss": 4.1367387771606445, + "step": 5880 + }, + { + "epoch": 81.68471615720524, + "grad_norm": 0.01647144928574562, + "learning_rate": 0.0006, + "loss": 4.096675395965576, + "step": 5881 + }, + { + "epoch": 81.69868995633188, + "grad_norm": 0.015978610143065453, + "learning_rate": 0.0006, + "loss": 4.232292175292969, + "step": 5882 + }, + { + "epoch": 81.71266375545852, + "grad_norm": 0.01510288193821907, + "learning_rate": 0.0006, + "loss": 4.274591445922852, + "step": 5883 + }, + { + "epoch": 81.72663755458515, + "grad_norm": 0.014942476525902748, + "learning_rate": 0.0006, + "loss": 4.161482810974121, + "step": 5884 + }, + { + "epoch": 81.74061135371178, + "grad_norm": 0.014595299027860165, + "learning_rate": 0.0006, + "loss": 4.204291343688965, + "step": 5885 + }, + { + "epoch": 81.75458515283843, + "grad_norm": 0.014952429570257664, + "learning_rate": 0.0006, + "loss": 4.242499351501465, + "step": 5886 + }, + { + "epoch": 81.76855895196506, + "grad_norm": 0.014761554077267647, + "learning_rate": 0.0006, + "loss": 4.20393180847168, + "step": 5887 + }, + { + "epoch": 81.78253275109171, + "grad_norm": 0.014749647118151188, + "learning_rate": 0.0006, + "loss": 4.227565765380859, + "step": 5888 + }, + { + "epoch": 81.79650655021834, + "grad_norm": 0.014384493231773376, + "learning_rate": 0.0006, + "loss": 4.282228946685791, + "step": 5889 + }, + { + "epoch": 81.81048034934499, + "grad_norm": 0.01260330155491829, + "learning_rate": 0.0006, + "loss": 4.2882866859436035, + "step": 5890 + }, + { + "epoch": 81.82445414847162, + "grad_norm": 0.014055311679840088, + "learning_rate": 0.0006, + "loss": 4.22351598739624, + "step": 5891 + }, + { + "epoch": 81.83842794759825, + "grad_norm": 0.013674065470695496, + "learning_rate": 0.0006, + "loss": 4.240716934204102, + "step": 5892 + }, + { + "epoch": 81.8524017467249, + "grad_norm": 0.01225196197628975, + "learning_rate": 0.0006, + "loss": 4.2958598136901855, + "step": 5893 + }, + { + "epoch": 81.86637554585153, + "grad_norm": 0.013538680970668793, + "learning_rate": 0.0006, + "loss": 4.24342679977417, + "step": 5894 + }, + { + "epoch": 81.88034934497817, + "grad_norm": 0.013151394203305244, + "learning_rate": 0.0006, + "loss": 4.081999778747559, + "step": 5895 + }, + { + "epoch": 81.8943231441048, + "grad_norm": 0.014260678552091122, + "learning_rate": 0.0006, + "loss": 4.156423568725586, + "step": 5896 + }, + { + "epoch": 81.90829694323143, + "grad_norm": 0.01329776830971241, + "learning_rate": 0.0006, + "loss": 4.1713995933532715, + "step": 5897 + }, + { + "epoch": 81.92227074235808, + "grad_norm": 0.014623276889324188, + "learning_rate": 0.0006, + "loss": 4.244891166687012, + "step": 5898 + }, + { + "epoch": 81.93624454148471, + "grad_norm": 0.015760324895381927, + "learning_rate": 0.0006, + "loss": 4.076528549194336, + "step": 5899 + }, + { + "epoch": 81.95021834061136, + "grad_norm": 0.014331766404211521, + "learning_rate": 0.0006, + "loss": 4.283037185668945, + "step": 5900 + }, + { + "epoch": 81.96419213973799, + "grad_norm": 0.012383842840790749, + "learning_rate": 0.0006, + "loss": 4.117308616638184, + "step": 5901 + }, + { + "epoch": 81.97816593886463, + "grad_norm": 0.01250660140067339, + "learning_rate": 0.0006, + "loss": 4.241250991821289, + "step": 5902 + }, + { + "epoch": 81.99213973799127, + "grad_norm": 0.012276900932192802, + "learning_rate": 0.0006, + "loss": 4.141091823577881, + "step": 5903 + }, + { + "epoch": 82.0, + "grad_norm": 0.01588737778365612, + "learning_rate": 0.0006, + "loss": 4.065532684326172, + "step": 5904 + }, + { + "epoch": 82.0, + "eval_loss": 4.612304210662842, + "eval_runtime": 57.5411, + "eval_samples_per_second": 42.439, + "eval_steps_per_second": 1.338, + "step": 5904 + }, + { + "epoch": 82.01397379912663, + "grad_norm": 0.017433544620871544, + "learning_rate": 0.0006, + "loss": 4.202497482299805, + "step": 5905 + }, + { + "epoch": 82.02794759825328, + "grad_norm": 0.017260415479540825, + "learning_rate": 0.0006, + "loss": 4.099949359893799, + "step": 5906 + }, + { + "epoch": 82.04192139737991, + "grad_norm": 0.01577908545732498, + "learning_rate": 0.0006, + "loss": 4.17806339263916, + "step": 5907 + }, + { + "epoch": 82.05589519650655, + "grad_norm": 0.016093818470835686, + "learning_rate": 0.0006, + "loss": 4.219491004943848, + "step": 5908 + }, + { + "epoch": 82.06986899563319, + "grad_norm": 0.015426302328705788, + "learning_rate": 0.0006, + "loss": 4.207982063293457, + "step": 5909 + }, + { + "epoch": 82.08384279475983, + "grad_norm": 0.01605442352592945, + "learning_rate": 0.0006, + "loss": 4.265942573547363, + "step": 5910 + }, + { + "epoch": 82.09781659388646, + "grad_norm": 0.01787225715816021, + "learning_rate": 0.0006, + "loss": 4.151535511016846, + "step": 5911 + }, + { + "epoch": 82.1117903930131, + "grad_norm": 0.01750164106488228, + "learning_rate": 0.0006, + "loss": 4.133655071258545, + "step": 5912 + }, + { + "epoch": 82.12576419213974, + "grad_norm": 0.01602957211434841, + "learning_rate": 0.0006, + "loss": 4.176972389221191, + "step": 5913 + }, + { + "epoch": 82.13973799126637, + "grad_norm": 0.01553812064230442, + "learning_rate": 0.0006, + "loss": 4.125692367553711, + "step": 5914 + }, + { + "epoch": 82.15371179039302, + "grad_norm": 0.013978508301079273, + "learning_rate": 0.0006, + "loss": 4.207271575927734, + "step": 5915 + }, + { + "epoch": 82.16768558951965, + "grad_norm": 0.013561118394136429, + "learning_rate": 0.0006, + "loss": 4.238930702209473, + "step": 5916 + }, + { + "epoch": 82.18165938864628, + "grad_norm": 0.013285507448017597, + "learning_rate": 0.0006, + "loss": 4.129464626312256, + "step": 5917 + }, + { + "epoch": 82.19563318777293, + "grad_norm": 0.013903839513659477, + "learning_rate": 0.0006, + "loss": 4.214109420776367, + "step": 5918 + }, + { + "epoch": 82.20960698689956, + "grad_norm": 0.014426837675273418, + "learning_rate": 0.0006, + "loss": 4.143196105957031, + "step": 5919 + }, + { + "epoch": 82.2235807860262, + "grad_norm": 0.015347165986895561, + "learning_rate": 0.0006, + "loss": 4.235439300537109, + "step": 5920 + }, + { + "epoch": 82.23755458515284, + "grad_norm": 0.014266039244830608, + "learning_rate": 0.0006, + "loss": 3.991617202758789, + "step": 5921 + }, + { + "epoch": 82.25152838427948, + "grad_norm": 0.013101593591272831, + "learning_rate": 0.0006, + "loss": 4.266017913818359, + "step": 5922 + }, + { + "epoch": 82.26550218340611, + "grad_norm": 0.013741881586611271, + "learning_rate": 0.0006, + "loss": 4.224879741668701, + "step": 5923 + }, + { + "epoch": 82.27947598253274, + "grad_norm": 0.014206945896148682, + "learning_rate": 0.0006, + "loss": 4.102272987365723, + "step": 5924 + }, + { + "epoch": 82.29344978165939, + "grad_norm": 0.013798723928630352, + "learning_rate": 0.0006, + "loss": 4.30418586730957, + "step": 5925 + }, + { + "epoch": 82.30742358078602, + "grad_norm": 0.015470122918486595, + "learning_rate": 0.0006, + "loss": 4.2513322830200195, + "step": 5926 + }, + { + "epoch": 82.32139737991267, + "grad_norm": 0.01655442826449871, + "learning_rate": 0.0006, + "loss": 4.223002910614014, + "step": 5927 + }, + { + "epoch": 82.3353711790393, + "grad_norm": 0.017170004546642303, + "learning_rate": 0.0006, + "loss": 4.253904819488525, + "step": 5928 + }, + { + "epoch": 82.34934497816595, + "grad_norm": 0.017595812678337097, + "learning_rate": 0.0006, + "loss": 4.184927940368652, + "step": 5929 + }, + { + "epoch": 82.36331877729258, + "grad_norm": 0.015073256567120552, + "learning_rate": 0.0006, + "loss": 4.202239036560059, + "step": 5930 + }, + { + "epoch": 82.37729257641921, + "grad_norm": 0.014310941100120544, + "learning_rate": 0.0006, + "loss": 4.124934673309326, + "step": 5931 + }, + { + "epoch": 82.39126637554585, + "grad_norm": 0.013671821914613247, + "learning_rate": 0.0006, + "loss": 4.138919353485107, + "step": 5932 + }, + { + "epoch": 82.40524017467249, + "grad_norm": 0.014103314839303493, + "learning_rate": 0.0006, + "loss": 4.211289405822754, + "step": 5933 + }, + { + "epoch": 82.41921397379913, + "grad_norm": 0.014272668398916721, + "learning_rate": 0.0006, + "loss": 4.06290340423584, + "step": 5934 + }, + { + "epoch": 82.43318777292576, + "grad_norm": 0.01387016475200653, + "learning_rate": 0.0006, + "loss": 4.101334571838379, + "step": 5935 + }, + { + "epoch": 82.44716157205241, + "grad_norm": 0.013318683952093124, + "learning_rate": 0.0006, + "loss": 4.206409931182861, + "step": 5936 + }, + { + "epoch": 82.46113537117904, + "grad_norm": 0.013781268149614334, + "learning_rate": 0.0006, + "loss": 4.189474105834961, + "step": 5937 + }, + { + "epoch": 82.47510917030567, + "grad_norm": 0.013195851817727089, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 5938 + }, + { + "epoch": 82.48908296943232, + "grad_norm": 0.014358163811266422, + "learning_rate": 0.0006, + "loss": 4.203262805938721, + "step": 5939 + }, + { + "epoch": 82.50305676855895, + "grad_norm": 0.01403636671602726, + "learning_rate": 0.0006, + "loss": 4.079258441925049, + "step": 5940 + }, + { + "epoch": 82.5170305676856, + "grad_norm": 0.013272535055875778, + "learning_rate": 0.0006, + "loss": 4.19197940826416, + "step": 5941 + }, + { + "epoch": 82.53100436681223, + "grad_norm": 0.013787264935672283, + "learning_rate": 0.0006, + "loss": 4.12995719909668, + "step": 5942 + }, + { + "epoch": 82.54497816593886, + "grad_norm": 0.015020057559013367, + "learning_rate": 0.0006, + "loss": 4.148890495300293, + "step": 5943 + }, + { + "epoch": 82.5589519650655, + "grad_norm": 0.014199157245457172, + "learning_rate": 0.0006, + "loss": 4.201516151428223, + "step": 5944 + }, + { + "epoch": 82.57292576419214, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.240894317626953, + "step": 5945 + }, + { + "epoch": 82.58689956331878, + "grad_norm": 0.014767560176551342, + "learning_rate": 0.0006, + "loss": 4.173320770263672, + "step": 5946 + }, + { + "epoch": 82.60087336244541, + "grad_norm": 0.01612078957259655, + "learning_rate": 0.0006, + "loss": 4.266448020935059, + "step": 5947 + }, + { + "epoch": 82.61484716157206, + "grad_norm": 0.016706543043255806, + "learning_rate": 0.0006, + "loss": 4.229398250579834, + "step": 5948 + }, + { + "epoch": 82.62882096069869, + "grad_norm": 0.015630796551704407, + "learning_rate": 0.0006, + "loss": 4.181282997131348, + "step": 5949 + }, + { + "epoch": 82.64279475982532, + "grad_norm": 0.013103844597935677, + "learning_rate": 0.0006, + "loss": 4.309617042541504, + "step": 5950 + }, + { + "epoch": 82.65676855895197, + "grad_norm": 0.016136541962623596, + "learning_rate": 0.0006, + "loss": 4.224159240722656, + "step": 5951 + }, + { + "epoch": 82.6707423580786, + "grad_norm": 0.016160814091563225, + "learning_rate": 0.0006, + "loss": 4.267970085144043, + "step": 5952 + }, + { + "epoch": 82.68471615720524, + "grad_norm": 0.01672680303454399, + "learning_rate": 0.0006, + "loss": 4.230506896972656, + "step": 5953 + }, + { + "epoch": 82.69868995633188, + "grad_norm": 0.0159586351364851, + "learning_rate": 0.0006, + "loss": 4.159542083740234, + "step": 5954 + }, + { + "epoch": 82.71266375545852, + "grad_norm": 0.01461219135671854, + "learning_rate": 0.0006, + "loss": 4.27863073348999, + "step": 5955 + }, + { + "epoch": 82.72663755458515, + "grad_norm": 0.015088613145053387, + "learning_rate": 0.0006, + "loss": 4.272418975830078, + "step": 5956 + }, + { + "epoch": 82.74061135371178, + "grad_norm": 0.01594499684870243, + "learning_rate": 0.0006, + "loss": 4.271326065063477, + "step": 5957 + }, + { + "epoch": 82.75458515283843, + "grad_norm": 0.016648447141051292, + "learning_rate": 0.0006, + "loss": 4.141478538513184, + "step": 5958 + }, + { + "epoch": 82.76855895196506, + "grad_norm": 0.01540316641330719, + "learning_rate": 0.0006, + "loss": 4.111814498901367, + "step": 5959 + }, + { + "epoch": 82.78253275109171, + "grad_norm": 0.01634364202618599, + "learning_rate": 0.0006, + "loss": 4.3113112449646, + "step": 5960 + }, + { + "epoch": 82.79650655021834, + "grad_norm": 0.014768977649509907, + "learning_rate": 0.0006, + "loss": 4.184217929840088, + "step": 5961 + }, + { + "epoch": 82.81048034934499, + "grad_norm": 0.013453755527734756, + "learning_rate": 0.0006, + "loss": 4.247729301452637, + "step": 5962 + }, + { + "epoch": 82.82445414847162, + "grad_norm": 0.014731907285749912, + "learning_rate": 0.0006, + "loss": 4.352839469909668, + "step": 5963 + }, + { + "epoch": 82.83842794759825, + "grad_norm": 0.016047345474362373, + "learning_rate": 0.0006, + "loss": 4.257283687591553, + "step": 5964 + }, + { + "epoch": 82.8524017467249, + "grad_norm": 0.017141170799732208, + "learning_rate": 0.0006, + "loss": 4.22850227355957, + "step": 5965 + }, + { + "epoch": 82.86637554585153, + "grad_norm": 0.01370396837592125, + "learning_rate": 0.0006, + "loss": 4.1169257164001465, + "step": 5966 + }, + { + "epoch": 82.88034934497817, + "grad_norm": 0.012060081586241722, + "learning_rate": 0.0006, + "loss": 4.195242404937744, + "step": 5967 + }, + { + "epoch": 82.8943231441048, + "grad_norm": 0.015528378076851368, + "learning_rate": 0.0006, + "loss": 4.294939994812012, + "step": 5968 + }, + { + "epoch": 82.90829694323143, + "grad_norm": 0.018151041120290756, + "learning_rate": 0.0006, + "loss": 4.137192726135254, + "step": 5969 + }, + { + "epoch": 82.92227074235808, + "grad_norm": 0.017895735800266266, + "learning_rate": 0.0006, + "loss": 4.174036026000977, + "step": 5970 + }, + { + "epoch": 82.93624454148471, + "grad_norm": 0.015610386617481709, + "learning_rate": 0.0006, + "loss": 4.230136871337891, + "step": 5971 + }, + { + "epoch": 82.95021834061136, + "grad_norm": 0.01339928898960352, + "learning_rate": 0.0006, + "loss": 4.199074745178223, + "step": 5972 + }, + { + "epoch": 82.96419213973799, + "grad_norm": 0.015454214997589588, + "learning_rate": 0.0006, + "loss": 4.216109752655029, + "step": 5973 + }, + { + "epoch": 82.97816593886463, + "grad_norm": 0.014978132210671902, + "learning_rate": 0.0006, + "loss": 4.131511688232422, + "step": 5974 + }, + { + "epoch": 82.99213973799127, + "grad_norm": 0.015611122362315655, + "learning_rate": 0.0006, + "loss": 4.246518135070801, + "step": 5975 + }, + { + "epoch": 83.0, + "grad_norm": 0.016682764515280724, + "learning_rate": 0.0006, + "loss": 4.110680103302002, + "step": 5976 + }, + { + "epoch": 83.0, + "eval_loss": 4.641707897186279, + "eval_runtime": 57.2799, + "eval_samples_per_second": 42.633, + "eval_steps_per_second": 1.344, + "step": 5976 + }, + { + "epoch": 83.01397379912663, + "grad_norm": 0.015007982030510902, + "learning_rate": 0.0006, + "loss": 4.100912094116211, + "step": 5977 + }, + { + "epoch": 83.02794759825328, + "grad_norm": 0.017609436064958572, + "learning_rate": 0.0006, + "loss": 4.233399868011475, + "step": 5978 + }, + { + "epoch": 83.04192139737991, + "grad_norm": 0.019817272201180458, + "learning_rate": 0.0006, + "loss": 4.191173076629639, + "step": 5979 + }, + { + "epoch": 83.05589519650655, + "grad_norm": 0.017752140760421753, + "learning_rate": 0.0006, + "loss": 4.044437885284424, + "step": 5980 + }, + { + "epoch": 83.06986899563319, + "grad_norm": 0.015482760034501553, + "learning_rate": 0.0006, + "loss": 4.319931983947754, + "step": 5981 + }, + { + "epoch": 83.08384279475983, + "grad_norm": 0.015153988264501095, + "learning_rate": 0.0006, + "loss": 4.095277309417725, + "step": 5982 + }, + { + "epoch": 83.09781659388646, + "grad_norm": 0.017025282606482506, + "learning_rate": 0.0006, + "loss": 4.235110759735107, + "step": 5983 + }, + { + "epoch": 83.1117903930131, + "grad_norm": 0.01566586270928383, + "learning_rate": 0.0006, + "loss": 4.1934685707092285, + "step": 5984 + }, + { + "epoch": 83.12576419213974, + "grad_norm": 0.016419263556599617, + "learning_rate": 0.0006, + "loss": 4.234419822692871, + "step": 5985 + }, + { + "epoch": 83.13973799126637, + "grad_norm": 0.017132675275206566, + "learning_rate": 0.0006, + "loss": 4.221447944641113, + "step": 5986 + }, + { + "epoch": 83.15371179039302, + "grad_norm": 0.016138629987835884, + "learning_rate": 0.0006, + "loss": 4.199830055236816, + "step": 5987 + }, + { + "epoch": 83.16768558951965, + "grad_norm": 0.016210174188017845, + "learning_rate": 0.0006, + "loss": 4.081717491149902, + "step": 5988 + }, + { + "epoch": 83.18165938864628, + "grad_norm": 0.016055308282375336, + "learning_rate": 0.0006, + "loss": 4.247655391693115, + "step": 5989 + }, + { + "epoch": 83.19563318777293, + "grad_norm": 0.016132952645421028, + "learning_rate": 0.0006, + "loss": 4.262535572052002, + "step": 5990 + }, + { + "epoch": 83.20960698689956, + "grad_norm": 0.01779935695230961, + "learning_rate": 0.0006, + "loss": 4.153120040893555, + "step": 5991 + }, + { + "epoch": 83.2235807860262, + "grad_norm": 0.018360646441578865, + "learning_rate": 0.0006, + "loss": 4.232644081115723, + "step": 5992 + }, + { + "epoch": 83.23755458515284, + "grad_norm": 0.019629675894975662, + "learning_rate": 0.0006, + "loss": 3.9758734703063965, + "step": 5993 + }, + { + "epoch": 83.25152838427948, + "grad_norm": 0.017819996923208237, + "learning_rate": 0.0006, + "loss": 4.196932315826416, + "step": 5994 + }, + { + "epoch": 83.26550218340611, + "grad_norm": 0.01557681243866682, + "learning_rate": 0.0006, + "loss": 4.122654914855957, + "step": 5995 + }, + { + "epoch": 83.27947598253274, + "grad_norm": 0.017858853563666344, + "learning_rate": 0.0006, + "loss": 4.185330867767334, + "step": 5996 + }, + { + "epoch": 83.29344978165939, + "grad_norm": 0.018065953627228737, + "learning_rate": 0.0006, + "loss": 4.218471050262451, + "step": 5997 + }, + { + "epoch": 83.30742358078602, + "grad_norm": 0.017247939482331276, + "learning_rate": 0.0006, + "loss": 4.127076148986816, + "step": 5998 + }, + { + "epoch": 83.32139737991267, + "grad_norm": 0.017047952860593796, + "learning_rate": 0.0006, + "loss": 4.131356716156006, + "step": 5999 + }, + { + "epoch": 83.3353711790393, + "grad_norm": 0.015905629843473434, + "learning_rate": 0.0006, + "loss": 4.2239789962768555, + "step": 6000 + }, + { + "epoch": 83.34934497816595, + "grad_norm": 0.016522347927093506, + "learning_rate": 0.0006, + "loss": 4.054897785186768, + "step": 6001 + }, + { + "epoch": 83.36331877729258, + "grad_norm": 0.01622571051120758, + "learning_rate": 0.0006, + "loss": 4.284210681915283, + "step": 6002 + }, + { + "epoch": 83.37729257641921, + "grad_norm": 0.01640467159450054, + "learning_rate": 0.0006, + "loss": 4.268840789794922, + "step": 6003 + }, + { + "epoch": 83.39126637554585, + "grad_norm": 0.015698859468102455, + "learning_rate": 0.0006, + "loss": 4.179793357849121, + "step": 6004 + }, + { + "epoch": 83.40524017467249, + "grad_norm": 0.016694528982043266, + "learning_rate": 0.0006, + "loss": 4.132297515869141, + "step": 6005 + }, + { + "epoch": 83.41921397379913, + "grad_norm": 0.015345900319516659, + "learning_rate": 0.0006, + "loss": 4.090163230895996, + "step": 6006 + }, + { + "epoch": 83.43318777292576, + "grad_norm": 0.014722758904099464, + "learning_rate": 0.0006, + "loss": 4.179166793823242, + "step": 6007 + }, + { + "epoch": 83.44716157205241, + "grad_norm": 0.014884887263178825, + "learning_rate": 0.0006, + "loss": 4.137659072875977, + "step": 6008 + }, + { + "epoch": 83.46113537117904, + "grad_norm": 0.014041902497410774, + "learning_rate": 0.0006, + "loss": 4.1202073097229, + "step": 6009 + }, + { + "epoch": 83.47510917030567, + "grad_norm": 0.012799333781003952, + "learning_rate": 0.0006, + "loss": 4.092607498168945, + "step": 6010 + }, + { + "epoch": 83.48908296943232, + "grad_norm": 0.013272707350552082, + "learning_rate": 0.0006, + "loss": 4.132715225219727, + "step": 6011 + }, + { + "epoch": 83.50305676855895, + "grad_norm": 0.012934371829032898, + "learning_rate": 0.0006, + "loss": 4.230278968811035, + "step": 6012 + }, + { + "epoch": 83.5170305676856, + "grad_norm": 0.0120006762444973, + "learning_rate": 0.0006, + "loss": 4.140068531036377, + "step": 6013 + }, + { + "epoch": 83.53100436681223, + "grad_norm": 0.013616700656712055, + "learning_rate": 0.0006, + "loss": 4.20107364654541, + "step": 6014 + }, + { + "epoch": 83.54497816593886, + "grad_norm": 0.013957851566374302, + "learning_rate": 0.0006, + "loss": 4.124945163726807, + "step": 6015 + }, + { + "epoch": 83.5589519650655, + "grad_norm": 0.012762055732309818, + "learning_rate": 0.0006, + "loss": 4.022052764892578, + "step": 6016 + }, + { + "epoch": 83.57292576419214, + "grad_norm": 0.013080576434731483, + "learning_rate": 0.0006, + "loss": 4.103252410888672, + "step": 6017 + }, + { + "epoch": 83.58689956331878, + "grad_norm": 0.013957500457763672, + "learning_rate": 0.0006, + "loss": 4.235089302062988, + "step": 6018 + }, + { + "epoch": 83.60087336244541, + "grad_norm": 0.014707164838910103, + "learning_rate": 0.0006, + "loss": 4.191570281982422, + "step": 6019 + }, + { + "epoch": 83.61484716157206, + "grad_norm": 0.012807869352400303, + "learning_rate": 0.0006, + "loss": 4.294543266296387, + "step": 6020 + }, + { + "epoch": 83.62882096069869, + "grad_norm": 0.012027994729578495, + "learning_rate": 0.0006, + "loss": 4.13849401473999, + "step": 6021 + }, + { + "epoch": 83.64279475982532, + "grad_norm": 0.012147724628448486, + "learning_rate": 0.0006, + "loss": 4.127481937408447, + "step": 6022 + }, + { + "epoch": 83.65676855895197, + "grad_norm": 0.013241392560303211, + "learning_rate": 0.0006, + "loss": 4.218972682952881, + "step": 6023 + }, + { + "epoch": 83.6707423580786, + "grad_norm": 0.014071648940443993, + "learning_rate": 0.0006, + "loss": 4.096724033355713, + "step": 6024 + }, + { + "epoch": 83.68471615720524, + "grad_norm": 0.015212797559797764, + "learning_rate": 0.0006, + "loss": 4.27097225189209, + "step": 6025 + }, + { + "epoch": 83.69868995633188, + "grad_norm": 0.017647385597229004, + "learning_rate": 0.0006, + "loss": 4.209835529327393, + "step": 6026 + }, + { + "epoch": 83.71266375545852, + "grad_norm": 0.01709647662937641, + "learning_rate": 0.0006, + "loss": 4.042106628417969, + "step": 6027 + }, + { + "epoch": 83.72663755458515, + "grad_norm": 0.014530927874147892, + "learning_rate": 0.0006, + "loss": 4.181731700897217, + "step": 6028 + }, + { + "epoch": 83.74061135371178, + "grad_norm": 0.015356711111962795, + "learning_rate": 0.0006, + "loss": 4.314183235168457, + "step": 6029 + }, + { + "epoch": 83.75458515283843, + "grad_norm": 0.015500029549002647, + "learning_rate": 0.0006, + "loss": 4.06776762008667, + "step": 6030 + }, + { + "epoch": 83.76855895196506, + "grad_norm": 0.015351332724094391, + "learning_rate": 0.0006, + "loss": 4.185141563415527, + "step": 6031 + }, + { + "epoch": 83.78253275109171, + "grad_norm": 0.016678154468536377, + "learning_rate": 0.0006, + "loss": 4.179529666900635, + "step": 6032 + }, + { + "epoch": 83.79650655021834, + "grad_norm": 0.0156265702098608, + "learning_rate": 0.0006, + "loss": 4.238584041595459, + "step": 6033 + }, + { + "epoch": 83.81048034934499, + "grad_norm": 0.016418596729636192, + "learning_rate": 0.0006, + "loss": 4.1593523025512695, + "step": 6034 + }, + { + "epoch": 83.82445414847162, + "grad_norm": 0.016113586723804474, + "learning_rate": 0.0006, + "loss": 4.157470226287842, + "step": 6035 + }, + { + "epoch": 83.83842794759825, + "grad_norm": 0.014434042386710644, + "learning_rate": 0.0006, + "loss": 4.196226119995117, + "step": 6036 + }, + { + "epoch": 83.8524017467249, + "grad_norm": 0.012647042982280254, + "learning_rate": 0.0006, + "loss": 4.130589485168457, + "step": 6037 + }, + { + "epoch": 83.86637554585153, + "grad_norm": 0.014412877149879932, + "learning_rate": 0.0006, + "loss": 4.154845237731934, + "step": 6038 + }, + { + "epoch": 83.88034934497817, + "grad_norm": 0.014222027733922005, + "learning_rate": 0.0006, + "loss": 4.189724922180176, + "step": 6039 + }, + { + "epoch": 83.8943231441048, + "grad_norm": 0.013084255158901215, + "learning_rate": 0.0006, + "loss": 4.303338527679443, + "step": 6040 + }, + { + "epoch": 83.90829694323143, + "grad_norm": 0.012659232132136822, + "learning_rate": 0.0006, + "loss": 4.08192253112793, + "step": 6041 + }, + { + "epoch": 83.92227074235808, + "grad_norm": 0.01246276218444109, + "learning_rate": 0.0006, + "loss": 4.283324718475342, + "step": 6042 + }, + { + "epoch": 83.93624454148471, + "grad_norm": 0.01340118795633316, + "learning_rate": 0.0006, + "loss": 4.117908477783203, + "step": 6043 + }, + { + "epoch": 83.95021834061136, + "grad_norm": 0.014496722258627415, + "learning_rate": 0.0006, + "loss": 4.110318660736084, + "step": 6044 + }, + { + "epoch": 83.96419213973799, + "grad_norm": 0.013348298147320747, + "learning_rate": 0.0006, + "loss": 4.185139179229736, + "step": 6045 + }, + { + "epoch": 83.97816593886463, + "grad_norm": 0.013294277712702751, + "learning_rate": 0.0006, + "loss": 4.271645545959473, + "step": 6046 + }, + { + "epoch": 83.99213973799127, + "grad_norm": 0.014431594870984554, + "learning_rate": 0.0006, + "loss": 4.161383152008057, + "step": 6047 + }, + { + "epoch": 84.0, + "grad_norm": 0.016268927603960037, + "learning_rate": 0.0006, + "loss": 4.178324222564697, + "step": 6048 + }, + { + "epoch": 84.0, + "eval_loss": 4.6307549476623535, + "eval_runtime": 56.8521, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 6048 + }, + { + "epoch": 84.01397379912663, + "grad_norm": 0.014864908531308174, + "learning_rate": 0.0006, + "loss": 4.121300220489502, + "step": 6049 + }, + { + "epoch": 84.02794759825328, + "grad_norm": 0.014863367192447186, + "learning_rate": 0.0006, + "loss": 4.1650543212890625, + "step": 6050 + }, + { + "epoch": 84.04192139737991, + "grad_norm": 0.014379646629095078, + "learning_rate": 0.0006, + "loss": 4.0201239585876465, + "step": 6051 + }, + { + "epoch": 84.05589519650655, + "grad_norm": 0.014871201477944851, + "learning_rate": 0.0006, + "loss": 4.0820698738098145, + "step": 6052 + }, + { + "epoch": 84.06986899563319, + "grad_norm": 0.013085179962217808, + "learning_rate": 0.0006, + "loss": 4.223984718322754, + "step": 6053 + }, + { + "epoch": 84.08384279475983, + "grad_norm": 0.012485949322581291, + "learning_rate": 0.0006, + "loss": 4.182206153869629, + "step": 6054 + }, + { + "epoch": 84.09781659388646, + "grad_norm": 0.013653291389346123, + "learning_rate": 0.0006, + "loss": 4.142557144165039, + "step": 6055 + }, + { + "epoch": 84.1117903930131, + "grad_norm": 0.013930532149970531, + "learning_rate": 0.0006, + "loss": 4.13972282409668, + "step": 6056 + }, + { + "epoch": 84.12576419213974, + "grad_norm": 0.01297379657626152, + "learning_rate": 0.0006, + "loss": 4.144664764404297, + "step": 6057 + }, + { + "epoch": 84.13973799126637, + "grad_norm": 0.013210492208600044, + "learning_rate": 0.0006, + "loss": 4.253764629364014, + "step": 6058 + }, + { + "epoch": 84.15371179039302, + "grad_norm": 0.014650176279246807, + "learning_rate": 0.0006, + "loss": 4.246866226196289, + "step": 6059 + }, + { + "epoch": 84.16768558951965, + "grad_norm": 0.01482292078435421, + "learning_rate": 0.0006, + "loss": 4.140187740325928, + "step": 6060 + }, + { + "epoch": 84.18165938864628, + "grad_norm": 0.01592552848160267, + "learning_rate": 0.0006, + "loss": 4.170705318450928, + "step": 6061 + }, + { + "epoch": 84.19563318777293, + "grad_norm": 0.017152156680822372, + "learning_rate": 0.0006, + "loss": 4.126092433929443, + "step": 6062 + }, + { + "epoch": 84.20960698689956, + "grad_norm": 0.015880905091762543, + "learning_rate": 0.0006, + "loss": 4.161938190460205, + "step": 6063 + }, + { + "epoch": 84.2235807860262, + "grad_norm": 0.015451471321284771, + "learning_rate": 0.0006, + "loss": 4.263461112976074, + "step": 6064 + }, + { + "epoch": 84.23755458515284, + "grad_norm": 0.014365996234118938, + "learning_rate": 0.0006, + "loss": 4.14488410949707, + "step": 6065 + }, + { + "epoch": 84.25152838427948, + "grad_norm": 0.016364749521017075, + "learning_rate": 0.0006, + "loss": 4.112489700317383, + "step": 6066 + }, + { + "epoch": 84.26550218340611, + "grad_norm": 0.01791754551231861, + "learning_rate": 0.0006, + "loss": 4.133513450622559, + "step": 6067 + }, + { + "epoch": 84.27947598253274, + "grad_norm": 0.016609208658337593, + "learning_rate": 0.0006, + "loss": 4.201829433441162, + "step": 6068 + }, + { + "epoch": 84.29344978165939, + "grad_norm": 0.017587820068001747, + "learning_rate": 0.0006, + "loss": 4.203151702880859, + "step": 6069 + }, + { + "epoch": 84.30742358078602, + "grad_norm": 0.017047742381691933, + "learning_rate": 0.0006, + "loss": 4.077008247375488, + "step": 6070 + }, + { + "epoch": 84.32139737991267, + "grad_norm": 0.01524646207690239, + "learning_rate": 0.0006, + "loss": 4.198038101196289, + "step": 6071 + }, + { + "epoch": 84.3353711790393, + "grad_norm": 0.014196421951055527, + "learning_rate": 0.0006, + "loss": 4.164245128631592, + "step": 6072 + }, + { + "epoch": 84.34934497816595, + "grad_norm": 0.016291776672005653, + "learning_rate": 0.0006, + "loss": 4.150152206420898, + "step": 6073 + }, + { + "epoch": 84.36331877729258, + "grad_norm": 0.015627920627593994, + "learning_rate": 0.0006, + "loss": 4.110006332397461, + "step": 6074 + }, + { + "epoch": 84.37729257641921, + "grad_norm": 0.013251487165689468, + "learning_rate": 0.0006, + "loss": 4.101463317871094, + "step": 6075 + }, + { + "epoch": 84.39126637554585, + "grad_norm": 0.013601605780422688, + "learning_rate": 0.0006, + "loss": 4.191105842590332, + "step": 6076 + }, + { + "epoch": 84.40524017467249, + "grad_norm": 0.014878741465508938, + "learning_rate": 0.0006, + "loss": 4.153046131134033, + "step": 6077 + }, + { + "epoch": 84.41921397379913, + "grad_norm": 0.014783897437155247, + "learning_rate": 0.0006, + "loss": 4.362000465393066, + "step": 6078 + }, + { + "epoch": 84.43318777292576, + "grad_norm": 0.014221332035958767, + "learning_rate": 0.0006, + "loss": 4.144364833831787, + "step": 6079 + }, + { + "epoch": 84.44716157205241, + "grad_norm": 0.014911034144461155, + "learning_rate": 0.0006, + "loss": 4.088857650756836, + "step": 6080 + }, + { + "epoch": 84.46113537117904, + "grad_norm": 0.014924770221114159, + "learning_rate": 0.0006, + "loss": 4.1677446365356445, + "step": 6081 + }, + { + "epoch": 84.47510917030567, + "grad_norm": 0.014394409023225307, + "learning_rate": 0.0006, + "loss": 4.087103843688965, + "step": 6082 + }, + { + "epoch": 84.48908296943232, + "grad_norm": 0.013926676474511623, + "learning_rate": 0.0006, + "loss": 3.9972851276397705, + "step": 6083 + }, + { + "epoch": 84.50305676855895, + "grad_norm": 0.013069954700767994, + "learning_rate": 0.0006, + "loss": 3.998237133026123, + "step": 6084 + }, + { + "epoch": 84.5170305676856, + "grad_norm": 0.012748011387884617, + "learning_rate": 0.0006, + "loss": 4.195858955383301, + "step": 6085 + }, + { + "epoch": 84.53100436681223, + "grad_norm": 0.012278077192604542, + "learning_rate": 0.0006, + "loss": 4.102165222167969, + "step": 6086 + }, + { + "epoch": 84.54497816593886, + "grad_norm": 0.014259163290262222, + "learning_rate": 0.0006, + "loss": 4.092750549316406, + "step": 6087 + }, + { + "epoch": 84.5589519650655, + "grad_norm": 0.014553683809936047, + "learning_rate": 0.0006, + "loss": 4.1412763595581055, + "step": 6088 + }, + { + "epoch": 84.57292576419214, + "grad_norm": 0.013605243526399136, + "learning_rate": 0.0006, + "loss": 4.192173480987549, + "step": 6089 + }, + { + "epoch": 84.58689956331878, + "grad_norm": 0.014259358868002892, + "learning_rate": 0.0006, + "loss": 4.356374740600586, + "step": 6090 + }, + { + "epoch": 84.60087336244541, + "grad_norm": 0.014166025444865227, + "learning_rate": 0.0006, + "loss": 4.165982246398926, + "step": 6091 + }, + { + "epoch": 84.61484716157206, + "grad_norm": 0.012198803946375847, + "learning_rate": 0.0006, + "loss": 4.287566184997559, + "step": 6092 + }, + { + "epoch": 84.62882096069869, + "grad_norm": 0.013794238679111004, + "learning_rate": 0.0006, + "loss": 4.199882984161377, + "step": 6093 + }, + { + "epoch": 84.64279475982532, + "grad_norm": 0.014757642522454262, + "learning_rate": 0.0006, + "loss": 4.116573333740234, + "step": 6094 + }, + { + "epoch": 84.65676855895197, + "grad_norm": 0.014591689221560955, + "learning_rate": 0.0006, + "loss": 4.240896224975586, + "step": 6095 + }, + { + "epoch": 84.6707423580786, + "grad_norm": 0.015217579901218414, + "learning_rate": 0.0006, + "loss": 4.257142066955566, + "step": 6096 + }, + { + "epoch": 84.68471615720524, + "grad_norm": 0.015371371991932392, + "learning_rate": 0.0006, + "loss": 4.181037425994873, + "step": 6097 + }, + { + "epoch": 84.69868995633188, + "grad_norm": 0.014412983320653439, + "learning_rate": 0.0006, + "loss": 4.2207794189453125, + "step": 6098 + }, + { + "epoch": 84.71266375545852, + "grad_norm": 0.013773671351373196, + "learning_rate": 0.0006, + "loss": 4.211123466491699, + "step": 6099 + }, + { + "epoch": 84.72663755458515, + "grad_norm": 0.012838046066462994, + "learning_rate": 0.0006, + "loss": 4.175459384918213, + "step": 6100 + }, + { + "epoch": 84.74061135371178, + "grad_norm": 0.014094929210841656, + "learning_rate": 0.0006, + "loss": 4.1724958419799805, + "step": 6101 + }, + { + "epoch": 84.75458515283843, + "grad_norm": 0.014573262073099613, + "learning_rate": 0.0006, + "loss": 4.15885066986084, + "step": 6102 + }, + { + "epoch": 84.76855895196506, + "grad_norm": 0.011964819394052029, + "learning_rate": 0.0006, + "loss": 3.9804892539978027, + "step": 6103 + }, + { + "epoch": 84.78253275109171, + "grad_norm": 0.014140539802610874, + "learning_rate": 0.0006, + "loss": 4.203708648681641, + "step": 6104 + }, + { + "epoch": 84.79650655021834, + "grad_norm": 0.016618940979242325, + "learning_rate": 0.0006, + "loss": 4.147684097290039, + "step": 6105 + }, + { + "epoch": 84.81048034934499, + "grad_norm": 0.015094385482370853, + "learning_rate": 0.0006, + "loss": 4.392223834991455, + "step": 6106 + }, + { + "epoch": 84.82445414847162, + "grad_norm": 0.014386717230081558, + "learning_rate": 0.0006, + "loss": 4.253332138061523, + "step": 6107 + }, + { + "epoch": 84.83842794759825, + "grad_norm": 0.013877755962312222, + "learning_rate": 0.0006, + "loss": 4.075056076049805, + "step": 6108 + }, + { + "epoch": 84.8524017467249, + "grad_norm": 0.014062155969440937, + "learning_rate": 0.0006, + "loss": 4.072048664093018, + "step": 6109 + }, + { + "epoch": 84.86637554585153, + "grad_norm": 0.014445619657635689, + "learning_rate": 0.0006, + "loss": 4.151832580566406, + "step": 6110 + }, + { + "epoch": 84.88034934497817, + "grad_norm": 0.015116363763809204, + "learning_rate": 0.0006, + "loss": 4.088900566101074, + "step": 6111 + }, + { + "epoch": 84.8943231441048, + "grad_norm": 0.014425474219024181, + "learning_rate": 0.0006, + "loss": 4.200483322143555, + "step": 6112 + }, + { + "epoch": 84.90829694323143, + "grad_norm": 0.01449573878198862, + "learning_rate": 0.0006, + "loss": 4.239742279052734, + "step": 6113 + }, + { + "epoch": 84.92227074235808, + "grad_norm": 0.015698261559009552, + "learning_rate": 0.0006, + "loss": 4.197391510009766, + "step": 6114 + }, + { + "epoch": 84.93624454148471, + "grad_norm": 0.014545733109116554, + "learning_rate": 0.0006, + "loss": 4.035508155822754, + "step": 6115 + }, + { + "epoch": 84.95021834061136, + "grad_norm": 0.012796483002603054, + "learning_rate": 0.0006, + "loss": 4.195248126983643, + "step": 6116 + }, + { + "epoch": 84.96419213973799, + "grad_norm": 0.01264485064893961, + "learning_rate": 0.0006, + "loss": 4.158118724822998, + "step": 6117 + }, + { + "epoch": 84.97816593886463, + "grad_norm": 0.013695276342332363, + "learning_rate": 0.0006, + "loss": 4.255288124084473, + "step": 6118 + }, + { + "epoch": 84.99213973799127, + "grad_norm": 0.013459564186632633, + "learning_rate": 0.0006, + "loss": 4.204586505889893, + "step": 6119 + }, + { + "epoch": 85.0, + "grad_norm": 0.014312652871012688, + "learning_rate": 0.0006, + "loss": 4.167879104614258, + "step": 6120 + }, + { + "epoch": 85.0, + "eval_loss": 4.608152389526367, + "eval_runtime": 57.1903, + "eval_samples_per_second": 42.7, + "eval_steps_per_second": 1.346, + "step": 6120 + }, + { + "epoch": 85.01397379912663, + "grad_norm": 0.013918368145823479, + "learning_rate": 0.0006, + "loss": 4.129991054534912, + "step": 6121 + }, + { + "epoch": 85.02794759825328, + "grad_norm": 0.01604384370148182, + "learning_rate": 0.0006, + "loss": 4.233926773071289, + "step": 6122 + }, + { + "epoch": 85.04192139737991, + "grad_norm": 0.016565652564167976, + "learning_rate": 0.0006, + "loss": 4.129604339599609, + "step": 6123 + }, + { + "epoch": 85.05589519650655, + "grad_norm": 0.0149751678109169, + "learning_rate": 0.0006, + "loss": 4.165839195251465, + "step": 6124 + }, + { + "epoch": 85.06986899563319, + "grad_norm": 0.013812461867928505, + "learning_rate": 0.0006, + "loss": 4.153792858123779, + "step": 6125 + }, + { + "epoch": 85.08384279475983, + "grad_norm": 0.014432745985686779, + "learning_rate": 0.0006, + "loss": 4.170895576477051, + "step": 6126 + }, + { + "epoch": 85.09781659388646, + "grad_norm": 0.015208868309855461, + "learning_rate": 0.0006, + "loss": 4.10639762878418, + "step": 6127 + }, + { + "epoch": 85.1117903930131, + "grad_norm": 0.013209272176027298, + "learning_rate": 0.0006, + "loss": 4.200364112854004, + "step": 6128 + }, + { + "epoch": 85.12576419213974, + "grad_norm": 0.014807002618908882, + "learning_rate": 0.0006, + "loss": 4.172666549682617, + "step": 6129 + }, + { + "epoch": 85.13973799126637, + "grad_norm": 0.015662945806980133, + "learning_rate": 0.0006, + "loss": 4.145219326019287, + "step": 6130 + }, + { + "epoch": 85.15371179039302, + "grad_norm": 0.014598497189581394, + "learning_rate": 0.0006, + "loss": 4.275618553161621, + "step": 6131 + }, + { + "epoch": 85.16768558951965, + "grad_norm": 0.013778576627373695, + "learning_rate": 0.0006, + "loss": 4.204716682434082, + "step": 6132 + }, + { + "epoch": 85.18165938864628, + "grad_norm": 0.015762265771627426, + "learning_rate": 0.0006, + "loss": 4.145524024963379, + "step": 6133 + }, + { + "epoch": 85.19563318777293, + "grad_norm": 0.015579195693135262, + "learning_rate": 0.0006, + "loss": 4.150625228881836, + "step": 6134 + }, + { + "epoch": 85.20960698689956, + "grad_norm": 0.016641216352581978, + "learning_rate": 0.0006, + "loss": 4.141819953918457, + "step": 6135 + }, + { + "epoch": 85.2235807860262, + "grad_norm": 0.01632271707057953, + "learning_rate": 0.0006, + "loss": 4.030620574951172, + "step": 6136 + }, + { + "epoch": 85.23755458515284, + "grad_norm": 0.015564918518066406, + "learning_rate": 0.0006, + "loss": 4.147753715515137, + "step": 6137 + }, + { + "epoch": 85.25152838427948, + "grad_norm": 0.0161129217594862, + "learning_rate": 0.0006, + "loss": 4.128524303436279, + "step": 6138 + }, + { + "epoch": 85.26550218340611, + "grad_norm": 0.017766550183296204, + "learning_rate": 0.0006, + "loss": 4.134856224060059, + "step": 6139 + }, + { + "epoch": 85.27947598253274, + "grad_norm": 0.01869054324924946, + "learning_rate": 0.0006, + "loss": 4.231565475463867, + "step": 6140 + }, + { + "epoch": 85.29344978165939, + "grad_norm": 0.01917010359466076, + "learning_rate": 0.0006, + "loss": 4.21175479888916, + "step": 6141 + }, + { + "epoch": 85.30742358078602, + "grad_norm": 0.018116485327482224, + "learning_rate": 0.0006, + "loss": 4.122644424438477, + "step": 6142 + }, + { + "epoch": 85.32139737991267, + "grad_norm": 0.015949107706546783, + "learning_rate": 0.0006, + "loss": 4.135458946228027, + "step": 6143 + }, + { + "epoch": 85.3353711790393, + "grad_norm": 0.014783683232963085, + "learning_rate": 0.0006, + "loss": 4.118994235992432, + "step": 6144 + }, + { + "epoch": 85.34934497816595, + "grad_norm": 0.015406692400574684, + "learning_rate": 0.0006, + "loss": 4.157268524169922, + "step": 6145 + }, + { + "epoch": 85.36331877729258, + "grad_norm": 0.013893205672502518, + "learning_rate": 0.0006, + "loss": 4.056166648864746, + "step": 6146 + }, + { + "epoch": 85.37729257641921, + "grad_norm": 0.014380923472344875, + "learning_rate": 0.0006, + "loss": 4.206905364990234, + "step": 6147 + }, + { + "epoch": 85.39126637554585, + "grad_norm": 0.015669258311390877, + "learning_rate": 0.0006, + "loss": 4.213881492614746, + "step": 6148 + }, + { + "epoch": 85.40524017467249, + "grad_norm": 0.01553428452461958, + "learning_rate": 0.0006, + "loss": 4.1814374923706055, + "step": 6149 + }, + { + "epoch": 85.41921397379913, + "grad_norm": 0.01697700284421444, + "learning_rate": 0.0006, + "loss": 4.083939552307129, + "step": 6150 + }, + { + "epoch": 85.43318777292576, + "grad_norm": 0.016639070585370064, + "learning_rate": 0.0006, + "loss": 4.279311180114746, + "step": 6151 + }, + { + "epoch": 85.44716157205241, + "grad_norm": 0.016285281628370285, + "learning_rate": 0.0006, + "loss": 4.244845390319824, + "step": 6152 + }, + { + "epoch": 85.46113537117904, + "grad_norm": 0.016071317717432976, + "learning_rate": 0.0006, + "loss": 4.089814186096191, + "step": 6153 + }, + { + "epoch": 85.47510917030567, + "grad_norm": 0.01640418916940689, + "learning_rate": 0.0006, + "loss": 4.168088912963867, + "step": 6154 + }, + { + "epoch": 85.48908296943232, + "grad_norm": 0.014781414531171322, + "learning_rate": 0.0006, + "loss": 4.196092128753662, + "step": 6155 + }, + { + "epoch": 85.50305676855895, + "grad_norm": 0.013393213972449303, + "learning_rate": 0.0006, + "loss": 4.193429946899414, + "step": 6156 + }, + { + "epoch": 85.5170305676856, + "grad_norm": 0.01467406190931797, + "learning_rate": 0.0006, + "loss": 4.216714859008789, + "step": 6157 + }, + { + "epoch": 85.53100436681223, + "grad_norm": 0.014950765296816826, + "learning_rate": 0.0006, + "loss": 4.093854904174805, + "step": 6158 + }, + { + "epoch": 85.54497816593886, + "grad_norm": 0.014655301347374916, + "learning_rate": 0.0006, + "loss": 4.09483528137207, + "step": 6159 + }, + { + "epoch": 85.5589519650655, + "grad_norm": 0.01425520796328783, + "learning_rate": 0.0006, + "loss": 4.090396881103516, + "step": 6160 + }, + { + "epoch": 85.57292576419214, + "grad_norm": 0.01403278298676014, + "learning_rate": 0.0006, + "loss": 4.088724136352539, + "step": 6161 + }, + { + "epoch": 85.58689956331878, + "grad_norm": 0.01364967692643404, + "learning_rate": 0.0006, + "loss": 4.095558166503906, + "step": 6162 + }, + { + "epoch": 85.60087336244541, + "grad_norm": 0.013085267506539822, + "learning_rate": 0.0006, + "loss": 4.163962364196777, + "step": 6163 + }, + { + "epoch": 85.61484716157206, + "grad_norm": 0.0136906448751688, + "learning_rate": 0.0006, + "loss": 4.197134017944336, + "step": 6164 + }, + { + "epoch": 85.62882096069869, + "grad_norm": 0.013472451828420162, + "learning_rate": 0.0006, + "loss": 4.1578369140625, + "step": 6165 + }, + { + "epoch": 85.64279475982532, + "grad_norm": 0.012329773046076298, + "learning_rate": 0.0006, + "loss": 4.231907844543457, + "step": 6166 + }, + { + "epoch": 85.65676855895197, + "grad_norm": 0.011977200396358967, + "learning_rate": 0.0006, + "loss": 4.1552958488464355, + "step": 6167 + }, + { + "epoch": 85.6707423580786, + "grad_norm": 0.013049956411123276, + "learning_rate": 0.0006, + "loss": 4.17852783203125, + "step": 6168 + }, + { + "epoch": 85.68471615720524, + "grad_norm": 0.011440741829574108, + "learning_rate": 0.0006, + "loss": 4.144646644592285, + "step": 6169 + }, + { + "epoch": 85.69868995633188, + "grad_norm": 0.012472936883568764, + "learning_rate": 0.0006, + "loss": 4.2306413650512695, + "step": 6170 + }, + { + "epoch": 85.71266375545852, + "grad_norm": 0.013015356846153736, + "learning_rate": 0.0006, + "loss": 4.193439960479736, + "step": 6171 + }, + { + "epoch": 85.72663755458515, + "grad_norm": 0.013832319527864456, + "learning_rate": 0.0006, + "loss": 4.18073034286499, + "step": 6172 + }, + { + "epoch": 85.74061135371178, + "grad_norm": 0.014613982290029526, + "learning_rate": 0.0006, + "loss": 4.223531246185303, + "step": 6173 + }, + { + "epoch": 85.75458515283843, + "grad_norm": 0.014209083281457424, + "learning_rate": 0.0006, + "loss": 4.142666816711426, + "step": 6174 + }, + { + "epoch": 85.76855895196506, + "grad_norm": 0.013396669179201126, + "learning_rate": 0.0006, + "loss": 4.185786247253418, + "step": 6175 + }, + { + "epoch": 85.78253275109171, + "grad_norm": 0.01364929135888815, + "learning_rate": 0.0006, + "loss": 4.125226974487305, + "step": 6176 + }, + { + "epoch": 85.79650655021834, + "grad_norm": 0.012561003677546978, + "learning_rate": 0.0006, + "loss": 4.084519386291504, + "step": 6177 + }, + { + "epoch": 85.81048034934499, + "grad_norm": 0.013508481904864311, + "learning_rate": 0.0006, + "loss": 4.193173885345459, + "step": 6178 + }, + { + "epoch": 85.82445414847162, + "grad_norm": 0.014154613018035889, + "learning_rate": 0.0006, + "loss": 4.257272720336914, + "step": 6179 + }, + { + "epoch": 85.83842794759825, + "grad_norm": 0.012478531338274479, + "learning_rate": 0.0006, + "loss": 4.161380767822266, + "step": 6180 + }, + { + "epoch": 85.8524017467249, + "grad_norm": 0.013468451797962189, + "learning_rate": 0.0006, + "loss": 4.121395111083984, + "step": 6181 + }, + { + "epoch": 85.86637554585153, + "grad_norm": 0.015155194327235222, + "learning_rate": 0.0006, + "loss": 4.242815971374512, + "step": 6182 + }, + { + "epoch": 85.88034934497817, + "grad_norm": 0.016223272308707237, + "learning_rate": 0.0006, + "loss": 4.176027297973633, + "step": 6183 + }, + { + "epoch": 85.8943231441048, + "grad_norm": 0.013741295784711838, + "learning_rate": 0.0006, + "loss": 4.131257057189941, + "step": 6184 + }, + { + "epoch": 85.90829694323143, + "grad_norm": 0.014982067979872227, + "learning_rate": 0.0006, + "loss": 4.237896919250488, + "step": 6185 + }, + { + "epoch": 85.92227074235808, + "grad_norm": 0.018197322264313698, + "learning_rate": 0.0006, + "loss": 4.154962539672852, + "step": 6186 + }, + { + "epoch": 85.93624454148471, + "grad_norm": 0.01971200481057167, + "learning_rate": 0.0006, + "loss": 4.16819429397583, + "step": 6187 + }, + { + "epoch": 85.95021834061136, + "grad_norm": 0.018008720129728317, + "learning_rate": 0.0006, + "loss": 4.146164417266846, + "step": 6188 + }, + { + "epoch": 85.96419213973799, + "grad_norm": 0.015859264880418777, + "learning_rate": 0.0006, + "loss": 4.208099365234375, + "step": 6189 + }, + { + "epoch": 85.97816593886463, + "grad_norm": 0.013846252113580704, + "learning_rate": 0.0006, + "loss": 4.1881184577941895, + "step": 6190 + }, + { + "epoch": 85.99213973799127, + "grad_norm": 0.018641939386725426, + "learning_rate": 0.0006, + "loss": 4.127632141113281, + "step": 6191 + }, + { + "epoch": 86.0, + "grad_norm": 0.02385227382183075, + "learning_rate": 0.0006, + "loss": 4.281900405883789, + "step": 6192 + }, + { + "epoch": 86.0, + "eval_loss": 4.625458717346191, + "eval_runtime": 57.0634, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6192 + }, + { + "epoch": 86.01397379912663, + "grad_norm": 0.019032124429941177, + "learning_rate": 0.0006, + "loss": 4.2163190841674805, + "step": 6193 + }, + { + "epoch": 86.02794759825328, + "grad_norm": 0.018049925565719604, + "learning_rate": 0.0006, + "loss": 4.1534037590026855, + "step": 6194 + }, + { + "epoch": 86.04192139737991, + "grad_norm": 0.019116360694169998, + "learning_rate": 0.0006, + "loss": 4.212411403656006, + "step": 6195 + }, + { + "epoch": 86.05589519650655, + "grad_norm": 0.018967336043715477, + "learning_rate": 0.0006, + "loss": 4.190918922424316, + "step": 6196 + }, + { + "epoch": 86.06986899563319, + "grad_norm": 0.01945611648261547, + "learning_rate": 0.0006, + "loss": 4.18459415435791, + "step": 6197 + }, + { + "epoch": 86.08384279475983, + "grad_norm": 0.018104076385498047, + "learning_rate": 0.0006, + "loss": 4.0942816734313965, + "step": 6198 + }, + { + "epoch": 86.09781659388646, + "grad_norm": 0.01815992221236229, + "learning_rate": 0.0006, + "loss": 4.011547565460205, + "step": 6199 + }, + { + "epoch": 86.1117903930131, + "grad_norm": 0.017574431374669075, + "learning_rate": 0.0006, + "loss": 4.051593780517578, + "step": 6200 + }, + { + "epoch": 86.12576419213974, + "grad_norm": 0.01516508124768734, + "learning_rate": 0.0006, + "loss": 4.239673614501953, + "step": 6201 + }, + { + "epoch": 86.13973799126637, + "grad_norm": 0.015689702704548836, + "learning_rate": 0.0006, + "loss": 4.1085734367370605, + "step": 6202 + }, + { + "epoch": 86.15371179039302, + "grad_norm": 0.01596900261938572, + "learning_rate": 0.0006, + "loss": 4.1509599685668945, + "step": 6203 + }, + { + "epoch": 86.16768558951965, + "grad_norm": 0.01550096832215786, + "learning_rate": 0.0006, + "loss": 4.160238265991211, + "step": 6204 + }, + { + "epoch": 86.18165938864628, + "grad_norm": 0.014601529575884342, + "learning_rate": 0.0006, + "loss": 4.029219150543213, + "step": 6205 + }, + { + "epoch": 86.19563318777293, + "grad_norm": 0.014006740413606167, + "learning_rate": 0.0006, + "loss": 4.117501258850098, + "step": 6206 + }, + { + "epoch": 86.20960698689956, + "grad_norm": 0.013191957026720047, + "learning_rate": 0.0006, + "loss": 4.252002716064453, + "step": 6207 + }, + { + "epoch": 86.2235807860262, + "grad_norm": 0.013031814247369766, + "learning_rate": 0.0006, + "loss": 4.209517478942871, + "step": 6208 + }, + { + "epoch": 86.23755458515284, + "grad_norm": 0.013605714775621891, + "learning_rate": 0.0006, + "loss": 4.176309585571289, + "step": 6209 + }, + { + "epoch": 86.25152838427948, + "grad_norm": 0.012753087095916271, + "learning_rate": 0.0006, + "loss": 4.097990036010742, + "step": 6210 + }, + { + "epoch": 86.26550218340611, + "grad_norm": 0.012940628454089165, + "learning_rate": 0.0006, + "loss": 4.085589408874512, + "step": 6211 + }, + { + "epoch": 86.27947598253274, + "grad_norm": 0.014286619611084461, + "learning_rate": 0.0006, + "loss": 4.199261665344238, + "step": 6212 + }, + { + "epoch": 86.29344978165939, + "grad_norm": 0.014643525704741478, + "learning_rate": 0.0006, + "loss": 4.23718786239624, + "step": 6213 + }, + { + "epoch": 86.30742358078602, + "grad_norm": 0.015182147733867168, + "learning_rate": 0.0006, + "loss": 4.219855308532715, + "step": 6214 + }, + { + "epoch": 86.32139737991267, + "grad_norm": 0.01530102826654911, + "learning_rate": 0.0006, + "loss": 4.190094947814941, + "step": 6215 + }, + { + "epoch": 86.3353711790393, + "grad_norm": 0.015492603182792664, + "learning_rate": 0.0006, + "loss": 4.153780460357666, + "step": 6216 + }, + { + "epoch": 86.34934497816595, + "grad_norm": 0.015339907258749008, + "learning_rate": 0.0006, + "loss": 4.182865619659424, + "step": 6217 + }, + { + "epoch": 86.36331877729258, + "grad_norm": 0.015576880425214767, + "learning_rate": 0.0006, + "loss": 4.196417331695557, + "step": 6218 + }, + { + "epoch": 86.37729257641921, + "grad_norm": 0.014805737882852554, + "learning_rate": 0.0006, + "loss": 4.265018463134766, + "step": 6219 + }, + { + "epoch": 86.39126637554585, + "grad_norm": 0.01597665622830391, + "learning_rate": 0.0006, + "loss": 4.2864837646484375, + "step": 6220 + }, + { + "epoch": 86.40524017467249, + "grad_norm": 0.01693643443286419, + "learning_rate": 0.0006, + "loss": 4.118192672729492, + "step": 6221 + }, + { + "epoch": 86.41921397379913, + "grad_norm": 0.016648951917886734, + "learning_rate": 0.0006, + "loss": 4.23712682723999, + "step": 6222 + }, + { + "epoch": 86.43318777292576, + "grad_norm": 0.014809766784310341, + "learning_rate": 0.0006, + "loss": 4.251391410827637, + "step": 6223 + }, + { + "epoch": 86.44716157205241, + "grad_norm": 0.014333373866975307, + "learning_rate": 0.0006, + "loss": 4.17243766784668, + "step": 6224 + }, + { + "epoch": 86.46113537117904, + "grad_norm": 0.013912711292505264, + "learning_rate": 0.0006, + "loss": 4.250836372375488, + "step": 6225 + }, + { + "epoch": 86.47510917030567, + "grad_norm": 0.014821341261267662, + "learning_rate": 0.0006, + "loss": 4.073995590209961, + "step": 6226 + }, + { + "epoch": 86.48908296943232, + "grad_norm": 0.016634922474622726, + "learning_rate": 0.0006, + "loss": 4.155867576599121, + "step": 6227 + }, + { + "epoch": 86.50305676855895, + "grad_norm": 0.01748695969581604, + "learning_rate": 0.0006, + "loss": 4.139344692230225, + "step": 6228 + }, + { + "epoch": 86.5170305676856, + "grad_norm": 0.01569363661110401, + "learning_rate": 0.0006, + "loss": 4.23923397064209, + "step": 6229 + }, + { + "epoch": 86.53100436681223, + "grad_norm": 0.014037121087312698, + "learning_rate": 0.0006, + "loss": 4.200334072113037, + "step": 6230 + }, + { + "epoch": 86.54497816593886, + "grad_norm": 0.01401279866695404, + "learning_rate": 0.0006, + "loss": 4.035246849060059, + "step": 6231 + }, + { + "epoch": 86.5589519650655, + "grad_norm": 0.014603731222450733, + "learning_rate": 0.0006, + "loss": 4.141385555267334, + "step": 6232 + }, + { + "epoch": 86.57292576419214, + "grad_norm": 0.015113704837858677, + "learning_rate": 0.0006, + "loss": 4.130824089050293, + "step": 6233 + }, + { + "epoch": 86.58689956331878, + "grad_norm": 0.01588696800172329, + "learning_rate": 0.0006, + "loss": 4.05495023727417, + "step": 6234 + }, + { + "epoch": 86.60087336244541, + "grad_norm": 0.01593640074133873, + "learning_rate": 0.0006, + "loss": 4.22935152053833, + "step": 6235 + }, + { + "epoch": 86.61484716157206, + "grad_norm": 0.014292445965111256, + "learning_rate": 0.0006, + "loss": 4.309839248657227, + "step": 6236 + }, + { + "epoch": 86.62882096069869, + "grad_norm": 0.015471610240638256, + "learning_rate": 0.0006, + "loss": 4.268278121948242, + "step": 6237 + }, + { + "epoch": 86.64279475982532, + "grad_norm": 0.016643544659018517, + "learning_rate": 0.0006, + "loss": 4.135109901428223, + "step": 6238 + }, + { + "epoch": 86.65676855895197, + "grad_norm": 0.014733029529452324, + "learning_rate": 0.0006, + "loss": 4.060197830200195, + "step": 6239 + }, + { + "epoch": 86.6707423580786, + "grad_norm": 0.014869216829538345, + "learning_rate": 0.0006, + "loss": 4.126917362213135, + "step": 6240 + }, + { + "epoch": 86.68471615720524, + "grad_norm": 0.01608203910291195, + "learning_rate": 0.0006, + "loss": 4.183716773986816, + "step": 6241 + }, + { + "epoch": 86.69868995633188, + "grad_norm": 0.016178175806999207, + "learning_rate": 0.0006, + "loss": 4.118875503540039, + "step": 6242 + }, + { + "epoch": 86.71266375545852, + "grad_norm": 0.014576597139239311, + "learning_rate": 0.0006, + "loss": 4.261541366577148, + "step": 6243 + }, + { + "epoch": 86.72663755458515, + "grad_norm": 0.014466420747339725, + "learning_rate": 0.0006, + "loss": 4.307284355163574, + "step": 6244 + }, + { + "epoch": 86.74061135371178, + "grad_norm": 0.014042644761502743, + "learning_rate": 0.0006, + "loss": 4.2729997634887695, + "step": 6245 + }, + { + "epoch": 86.75458515283843, + "grad_norm": 0.014589524827897549, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6246 + }, + { + "epoch": 86.76855895196506, + "grad_norm": 0.018765483051538467, + "learning_rate": 0.0006, + "loss": 4.264769554138184, + "step": 6247 + }, + { + "epoch": 86.78253275109171, + "grad_norm": 0.019397731870412827, + "learning_rate": 0.0006, + "loss": 4.183379173278809, + "step": 6248 + }, + { + "epoch": 86.79650655021834, + "grad_norm": 0.017525402829051018, + "learning_rate": 0.0006, + "loss": 4.1494245529174805, + "step": 6249 + }, + { + "epoch": 86.81048034934499, + "grad_norm": 0.016310200095176697, + "learning_rate": 0.0006, + "loss": 4.149104118347168, + "step": 6250 + }, + { + "epoch": 86.82445414847162, + "grad_norm": 0.01506423857063055, + "learning_rate": 0.0006, + "loss": 4.22091817855835, + "step": 6251 + }, + { + "epoch": 86.83842794759825, + "grad_norm": 0.014200284145772457, + "learning_rate": 0.0006, + "loss": 4.177569389343262, + "step": 6252 + }, + { + "epoch": 86.8524017467249, + "grad_norm": 0.015016552060842514, + "learning_rate": 0.0006, + "loss": 4.089859962463379, + "step": 6253 + }, + { + "epoch": 86.86637554585153, + "grad_norm": 0.016054971143603325, + "learning_rate": 0.0006, + "loss": 4.157166004180908, + "step": 6254 + }, + { + "epoch": 86.88034934497817, + "grad_norm": 0.014388521201908588, + "learning_rate": 0.0006, + "loss": 4.1437578201293945, + "step": 6255 + }, + { + "epoch": 86.8943231441048, + "grad_norm": 0.015705369412899017, + "learning_rate": 0.0006, + "loss": 4.045689582824707, + "step": 6256 + }, + { + "epoch": 86.90829694323143, + "grad_norm": 0.014192395843565464, + "learning_rate": 0.0006, + "loss": 4.098793029785156, + "step": 6257 + }, + { + "epoch": 86.92227074235808, + "grad_norm": 0.013751034624874592, + "learning_rate": 0.0006, + "loss": 3.989920139312744, + "step": 6258 + }, + { + "epoch": 86.93624454148471, + "grad_norm": 0.01582193188369274, + "learning_rate": 0.0006, + "loss": 4.283449172973633, + "step": 6259 + }, + { + "epoch": 86.95021834061136, + "grad_norm": 0.01476013008505106, + "learning_rate": 0.0006, + "loss": 4.221401691436768, + "step": 6260 + }, + { + "epoch": 86.96419213973799, + "grad_norm": 0.01346337515860796, + "learning_rate": 0.0006, + "loss": 4.05896520614624, + "step": 6261 + }, + { + "epoch": 86.97816593886463, + "grad_norm": 0.012899307534098625, + "learning_rate": 0.0006, + "loss": 4.159560203552246, + "step": 6262 + }, + { + "epoch": 86.99213973799127, + "grad_norm": 0.013150574639439583, + "learning_rate": 0.0006, + "loss": 4.217330455780029, + "step": 6263 + }, + { + "epoch": 87.0, + "grad_norm": 0.014940370805561543, + "learning_rate": 0.0006, + "loss": 4.1611833572387695, + "step": 6264 + }, + { + "epoch": 87.0, + "eval_loss": 4.666043758392334, + "eval_runtime": 56.5978, + "eval_samples_per_second": 43.147, + "eval_steps_per_second": 1.36, + "step": 6264 + }, + { + "epoch": 87.01397379912663, + "grad_norm": 0.013721180148422718, + "learning_rate": 0.0006, + "loss": 4.142454147338867, + "step": 6265 + }, + { + "epoch": 87.02794759825328, + "grad_norm": 0.014805255457758904, + "learning_rate": 0.0006, + "loss": 4.121677875518799, + "step": 6266 + }, + { + "epoch": 87.04192139737991, + "grad_norm": 0.015576514415442944, + "learning_rate": 0.0006, + "loss": 4.1188435554504395, + "step": 6267 + }, + { + "epoch": 87.05589519650655, + "grad_norm": 0.01605502888560295, + "learning_rate": 0.0006, + "loss": 4.255428314208984, + "step": 6268 + }, + { + "epoch": 87.06986899563319, + "grad_norm": 0.014933343045413494, + "learning_rate": 0.0006, + "loss": 4.122596740722656, + "step": 6269 + }, + { + "epoch": 87.08384279475983, + "grad_norm": 0.014630894176661968, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 6270 + }, + { + "epoch": 87.09781659388646, + "grad_norm": 0.014211163856089115, + "learning_rate": 0.0006, + "loss": 4.058030128479004, + "step": 6271 + }, + { + "epoch": 87.1117903930131, + "grad_norm": 0.014060168527066708, + "learning_rate": 0.0006, + "loss": 4.146586894989014, + "step": 6272 + }, + { + "epoch": 87.12576419213974, + "grad_norm": 0.014740100130438805, + "learning_rate": 0.0006, + "loss": 4.077739715576172, + "step": 6273 + }, + { + "epoch": 87.13973799126637, + "grad_norm": 0.013970664702355862, + "learning_rate": 0.0006, + "loss": 4.038252830505371, + "step": 6274 + }, + { + "epoch": 87.15371179039302, + "grad_norm": 0.014399951323866844, + "learning_rate": 0.0006, + "loss": 4.143700122833252, + "step": 6275 + }, + { + "epoch": 87.16768558951965, + "grad_norm": 0.01496410183608532, + "learning_rate": 0.0006, + "loss": 4.105609893798828, + "step": 6276 + }, + { + "epoch": 87.18165938864628, + "grad_norm": 0.014287355355918407, + "learning_rate": 0.0006, + "loss": 4.173645973205566, + "step": 6277 + }, + { + "epoch": 87.19563318777293, + "grad_norm": 0.0141292130574584, + "learning_rate": 0.0006, + "loss": 4.068511486053467, + "step": 6278 + }, + { + "epoch": 87.20960698689956, + "grad_norm": 0.01393523346632719, + "learning_rate": 0.0006, + "loss": 4.060687065124512, + "step": 6279 + }, + { + "epoch": 87.2235807860262, + "grad_norm": 0.01431021187454462, + "learning_rate": 0.0006, + "loss": 4.219209671020508, + "step": 6280 + }, + { + "epoch": 87.23755458515284, + "grad_norm": 0.014856027439236641, + "learning_rate": 0.0006, + "loss": 4.146749973297119, + "step": 6281 + }, + { + "epoch": 87.25152838427948, + "grad_norm": 0.013677539303898811, + "learning_rate": 0.0006, + "loss": 3.9859797954559326, + "step": 6282 + }, + { + "epoch": 87.26550218340611, + "grad_norm": 0.014078967273235321, + "learning_rate": 0.0006, + "loss": 4.129668235778809, + "step": 6283 + }, + { + "epoch": 87.27947598253274, + "grad_norm": 0.015564394183456898, + "learning_rate": 0.0006, + "loss": 4.0007829666137695, + "step": 6284 + }, + { + "epoch": 87.29344978165939, + "grad_norm": 0.015077749267220497, + "learning_rate": 0.0006, + "loss": 4.112264156341553, + "step": 6285 + }, + { + "epoch": 87.30742358078602, + "grad_norm": 0.015341218560934067, + "learning_rate": 0.0006, + "loss": 4.112249374389648, + "step": 6286 + }, + { + "epoch": 87.32139737991267, + "grad_norm": 0.016399366781115532, + "learning_rate": 0.0006, + "loss": 4.176196098327637, + "step": 6287 + }, + { + "epoch": 87.3353711790393, + "grad_norm": 0.01599763333797455, + "learning_rate": 0.0006, + "loss": 4.163441181182861, + "step": 6288 + }, + { + "epoch": 87.34934497816595, + "grad_norm": 0.015439776703715324, + "learning_rate": 0.0006, + "loss": 4.11902379989624, + "step": 6289 + }, + { + "epoch": 87.36331877729258, + "grad_norm": 0.015909943729639053, + "learning_rate": 0.0006, + "loss": 4.16621732711792, + "step": 6290 + }, + { + "epoch": 87.37729257641921, + "grad_norm": 0.016710078343749046, + "learning_rate": 0.0006, + "loss": 4.044960975646973, + "step": 6291 + }, + { + "epoch": 87.39126637554585, + "grad_norm": 0.015883933752775192, + "learning_rate": 0.0006, + "loss": 4.065184593200684, + "step": 6292 + }, + { + "epoch": 87.40524017467249, + "grad_norm": 0.013912747614085674, + "learning_rate": 0.0006, + "loss": 4.102450370788574, + "step": 6293 + }, + { + "epoch": 87.41921397379913, + "grad_norm": 0.013170420192182064, + "learning_rate": 0.0006, + "loss": 4.234364032745361, + "step": 6294 + }, + { + "epoch": 87.43318777292576, + "grad_norm": 0.014043302275240421, + "learning_rate": 0.0006, + "loss": 4.129565238952637, + "step": 6295 + }, + { + "epoch": 87.44716157205241, + "grad_norm": 0.015010139904916286, + "learning_rate": 0.0006, + "loss": 4.166678428649902, + "step": 6296 + }, + { + "epoch": 87.46113537117904, + "grad_norm": 0.014721784740686417, + "learning_rate": 0.0006, + "loss": 4.182618618011475, + "step": 6297 + }, + { + "epoch": 87.47510917030567, + "grad_norm": 0.015713468194007874, + "learning_rate": 0.0006, + "loss": 4.103855133056641, + "step": 6298 + }, + { + "epoch": 87.48908296943232, + "grad_norm": 0.017204247415065765, + "learning_rate": 0.0006, + "loss": 4.140937328338623, + "step": 6299 + }, + { + "epoch": 87.50305676855895, + "grad_norm": 0.016856668516993523, + "learning_rate": 0.0006, + "loss": 4.147859573364258, + "step": 6300 + }, + { + "epoch": 87.5170305676856, + "grad_norm": 0.017205018550157547, + "learning_rate": 0.0006, + "loss": 4.1667327880859375, + "step": 6301 + }, + { + "epoch": 87.53100436681223, + "grad_norm": 0.014821539632976055, + "learning_rate": 0.0006, + "loss": 4.089935302734375, + "step": 6302 + }, + { + "epoch": 87.54497816593886, + "grad_norm": 0.01446484588086605, + "learning_rate": 0.0006, + "loss": 4.243657112121582, + "step": 6303 + }, + { + "epoch": 87.5589519650655, + "grad_norm": 0.014374484308063984, + "learning_rate": 0.0006, + "loss": 4.14356803894043, + "step": 6304 + }, + { + "epoch": 87.57292576419214, + "grad_norm": 0.01420109998434782, + "learning_rate": 0.0006, + "loss": 4.1273651123046875, + "step": 6305 + }, + { + "epoch": 87.58689956331878, + "grad_norm": 0.014810388907790184, + "learning_rate": 0.0006, + "loss": 4.222290515899658, + "step": 6306 + }, + { + "epoch": 87.60087336244541, + "grad_norm": 0.015826983377337456, + "learning_rate": 0.0006, + "loss": 4.137996196746826, + "step": 6307 + }, + { + "epoch": 87.61484716157206, + "grad_norm": 0.014457454904913902, + "learning_rate": 0.0006, + "loss": 4.0780744552612305, + "step": 6308 + }, + { + "epoch": 87.62882096069869, + "grad_norm": 0.014458604156970978, + "learning_rate": 0.0006, + "loss": 4.1287922859191895, + "step": 6309 + }, + { + "epoch": 87.64279475982532, + "grad_norm": 0.015282071195542812, + "learning_rate": 0.0006, + "loss": 4.189508438110352, + "step": 6310 + }, + { + "epoch": 87.65676855895197, + "grad_norm": 0.017536642029881477, + "learning_rate": 0.0006, + "loss": 4.070173263549805, + "step": 6311 + }, + { + "epoch": 87.6707423580786, + "grad_norm": 0.017303580418229103, + "learning_rate": 0.0006, + "loss": 4.119133949279785, + "step": 6312 + }, + { + "epoch": 87.68471615720524, + "grad_norm": 0.016148580238223076, + "learning_rate": 0.0006, + "loss": 3.994354486465454, + "step": 6313 + }, + { + "epoch": 87.69868995633188, + "grad_norm": 0.01706942543387413, + "learning_rate": 0.0006, + "loss": 4.183113098144531, + "step": 6314 + }, + { + "epoch": 87.71266375545852, + "grad_norm": 0.016971010714769363, + "learning_rate": 0.0006, + "loss": 4.104086875915527, + "step": 6315 + }, + { + "epoch": 87.72663755458515, + "grad_norm": 0.015516847372055054, + "learning_rate": 0.0006, + "loss": 4.2409772872924805, + "step": 6316 + }, + { + "epoch": 87.74061135371178, + "grad_norm": 0.015352269634604454, + "learning_rate": 0.0006, + "loss": 4.026427268981934, + "step": 6317 + }, + { + "epoch": 87.75458515283843, + "grad_norm": 0.017702415585517883, + "learning_rate": 0.0006, + "loss": 4.177840232849121, + "step": 6318 + }, + { + "epoch": 87.76855895196506, + "grad_norm": 0.01629541628062725, + "learning_rate": 0.0006, + "loss": 4.229039192199707, + "step": 6319 + }, + { + "epoch": 87.78253275109171, + "grad_norm": 0.01503711100667715, + "learning_rate": 0.0006, + "loss": 4.214350700378418, + "step": 6320 + }, + { + "epoch": 87.79650655021834, + "grad_norm": 0.014913883060216904, + "learning_rate": 0.0006, + "loss": 4.168880939483643, + "step": 6321 + }, + { + "epoch": 87.81048034934499, + "grad_norm": 0.014513909816741943, + "learning_rate": 0.0006, + "loss": 4.162267684936523, + "step": 6322 + }, + { + "epoch": 87.82445414847162, + "grad_norm": 0.014635873027145863, + "learning_rate": 0.0006, + "loss": 4.220067977905273, + "step": 6323 + }, + { + "epoch": 87.83842794759825, + "grad_norm": 0.017102796584367752, + "learning_rate": 0.0006, + "loss": 4.1493377685546875, + "step": 6324 + }, + { + "epoch": 87.8524017467249, + "grad_norm": 0.015533884055912495, + "learning_rate": 0.0006, + "loss": 4.0646209716796875, + "step": 6325 + }, + { + "epoch": 87.86637554585153, + "grad_norm": 0.014617550186812878, + "learning_rate": 0.0006, + "loss": 4.1390533447265625, + "step": 6326 + }, + { + "epoch": 87.88034934497817, + "grad_norm": 0.014142967760562897, + "learning_rate": 0.0006, + "loss": 4.133934020996094, + "step": 6327 + }, + { + "epoch": 87.8943231441048, + "grad_norm": 0.013182433322072029, + "learning_rate": 0.0006, + "loss": 4.125327110290527, + "step": 6328 + }, + { + "epoch": 87.90829694323143, + "grad_norm": 0.013928229920566082, + "learning_rate": 0.0006, + "loss": 4.142777442932129, + "step": 6329 + }, + { + "epoch": 87.92227074235808, + "grad_norm": 0.015337013639509678, + "learning_rate": 0.0006, + "loss": 4.140210151672363, + "step": 6330 + }, + { + "epoch": 87.93624454148471, + "grad_norm": 0.014873536303639412, + "learning_rate": 0.0006, + "loss": 4.102636337280273, + "step": 6331 + }, + { + "epoch": 87.95021834061136, + "grad_norm": 0.01387431938201189, + "learning_rate": 0.0006, + "loss": 4.149059295654297, + "step": 6332 + }, + { + "epoch": 87.96419213973799, + "grad_norm": 0.013234943151473999, + "learning_rate": 0.0006, + "loss": 4.1564459800720215, + "step": 6333 + }, + { + "epoch": 87.97816593886463, + "grad_norm": 0.013781017623841763, + "learning_rate": 0.0006, + "loss": 4.189839839935303, + "step": 6334 + }, + { + "epoch": 87.99213973799127, + "grad_norm": 0.013733026571571827, + "learning_rate": 0.0006, + "loss": 4.29277229309082, + "step": 6335 + }, + { + "epoch": 88.0, + "grad_norm": 0.014752327464520931, + "learning_rate": 0.0006, + "loss": 4.163867950439453, + "step": 6336 + }, + { + "epoch": 88.0, + "eval_loss": 4.535566329956055, + "eval_runtime": 56.4677, + "eval_samples_per_second": 43.246, + "eval_steps_per_second": 1.364, + "step": 6336 + }, + { + "epoch": 88.01397379912663, + "grad_norm": 0.013958732597529888, + "learning_rate": 0.0006, + "loss": 4.200681209564209, + "step": 6337 + }, + { + "epoch": 88.02794759825328, + "grad_norm": 0.015283545479178429, + "learning_rate": 0.0006, + "loss": 4.1736016273498535, + "step": 6338 + }, + { + "epoch": 88.04192139737991, + "grad_norm": 0.016192059963941574, + "learning_rate": 0.0006, + "loss": 4.040589809417725, + "step": 6339 + }, + { + "epoch": 88.05589519650655, + "grad_norm": 0.018096154555678368, + "learning_rate": 0.0006, + "loss": 4.146696090698242, + "step": 6340 + }, + { + "epoch": 88.06986899563319, + "grad_norm": 0.01797664538025856, + "learning_rate": 0.0006, + "loss": 4.007248878479004, + "step": 6341 + }, + { + "epoch": 88.08384279475983, + "grad_norm": 0.016275612637400627, + "learning_rate": 0.0006, + "loss": 4.193206787109375, + "step": 6342 + }, + { + "epoch": 88.09781659388646, + "grad_norm": 0.014210258610546589, + "learning_rate": 0.0006, + "loss": 4.064548492431641, + "step": 6343 + }, + { + "epoch": 88.1117903930131, + "grad_norm": 0.016085151582956314, + "learning_rate": 0.0006, + "loss": 4.003668785095215, + "step": 6344 + }, + { + "epoch": 88.12576419213974, + "grad_norm": 0.01930483989417553, + "learning_rate": 0.0006, + "loss": 4.236578941345215, + "step": 6345 + }, + { + "epoch": 88.13973799126637, + "grad_norm": 0.017140906304121017, + "learning_rate": 0.0006, + "loss": 3.9593005180358887, + "step": 6346 + }, + { + "epoch": 88.15371179039302, + "grad_norm": 0.01521560549736023, + "learning_rate": 0.0006, + "loss": 4.137644290924072, + "step": 6347 + }, + { + "epoch": 88.16768558951965, + "grad_norm": 0.015788141638040543, + "learning_rate": 0.0006, + "loss": 4.124845027923584, + "step": 6348 + }, + { + "epoch": 88.18165938864628, + "grad_norm": 0.01784597896039486, + "learning_rate": 0.0006, + "loss": 4.139750957489014, + "step": 6349 + }, + { + "epoch": 88.19563318777293, + "grad_norm": 0.016266558319330215, + "learning_rate": 0.0006, + "loss": 4.074851989746094, + "step": 6350 + }, + { + "epoch": 88.20960698689956, + "grad_norm": 0.014024483971297741, + "learning_rate": 0.0006, + "loss": 4.164689064025879, + "step": 6351 + }, + { + "epoch": 88.2235807860262, + "grad_norm": 0.015694420784711838, + "learning_rate": 0.0006, + "loss": 4.219979286193848, + "step": 6352 + }, + { + "epoch": 88.23755458515284, + "grad_norm": 0.014897859655320644, + "learning_rate": 0.0006, + "loss": 4.050738334655762, + "step": 6353 + }, + { + "epoch": 88.25152838427948, + "grad_norm": 0.014827374368906021, + "learning_rate": 0.0006, + "loss": 4.137135028839111, + "step": 6354 + }, + { + "epoch": 88.26550218340611, + "grad_norm": 0.01454879529774189, + "learning_rate": 0.0006, + "loss": 3.9980897903442383, + "step": 6355 + }, + { + "epoch": 88.27947598253274, + "grad_norm": 0.015623947605490685, + "learning_rate": 0.0006, + "loss": 4.226551055908203, + "step": 6356 + }, + { + "epoch": 88.29344978165939, + "grad_norm": 0.015527035109698772, + "learning_rate": 0.0006, + "loss": 4.11053466796875, + "step": 6357 + }, + { + "epoch": 88.30742358078602, + "grad_norm": 0.015242574736475945, + "learning_rate": 0.0006, + "loss": 4.105966567993164, + "step": 6358 + }, + { + "epoch": 88.32139737991267, + "grad_norm": 0.01375538855791092, + "learning_rate": 0.0006, + "loss": 4.148453712463379, + "step": 6359 + }, + { + "epoch": 88.3353711790393, + "grad_norm": 0.01595226116478443, + "learning_rate": 0.0006, + "loss": 4.095836162567139, + "step": 6360 + }, + { + "epoch": 88.34934497816595, + "grad_norm": 0.015252740122377872, + "learning_rate": 0.0006, + "loss": 4.136335849761963, + "step": 6361 + }, + { + "epoch": 88.36331877729258, + "grad_norm": 0.014405866153538227, + "learning_rate": 0.0006, + "loss": 4.084691047668457, + "step": 6362 + }, + { + "epoch": 88.37729257641921, + "grad_norm": 0.014755490235984325, + "learning_rate": 0.0006, + "loss": 4.095475673675537, + "step": 6363 + }, + { + "epoch": 88.39126637554585, + "grad_norm": 0.015339708887040615, + "learning_rate": 0.0006, + "loss": 4.163136959075928, + "step": 6364 + }, + { + "epoch": 88.40524017467249, + "grad_norm": 0.014667164534330368, + "learning_rate": 0.0006, + "loss": 4.062265396118164, + "step": 6365 + }, + { + "epoch": 88.41921397379913, + "grad_norm": 0.014588423073291779, + "learning_rate": 0.0006, + "loss": 4.13359260559082, + "step": 6366 + }, + { + "epoch": 88.43318777292576, + "grad_norm": 0.014005377888679504, + "learning_rate": 0.0006, + "loss": 4.219711780548096, + "step": 6367 + }, + { + "epoch": 88.44716157205241, + "grad_norm": 0.012769930064678192, + "learning_rate": 0.0006, + "loss": 4.114658355712891, + "step": 6368 + }, + { + "epoch": 88.46113537117904, + "grad_norm": 0.013632693327963352, + "learning_rate": 0.0006, + "loss": 4.0675225257873535, + "step": 6369 + }, + { + "epoch": 88.47510917030567, + "grad_norm": 0.013829804956912994, + "learning_rate": 0.0006, + "loss": 4.1125969886779785, + "step": 6370 + }, + { + "epoch": 88.48908296943232, + "grad_norm": 0.013882109895348549, + "learning_rate": 0.0006, + "loss": 4.1627092361450195, + "step": 6371 + }, + { + "epoch": 88.50305676855895, + "grad_norm": 0.014023828320205212, + "learning_rate": 0.0006, + "loss": 4.01038932800293, + "step": 6372 + }, + { + "epoch": 88.5170305676856, + "grad_norm": 0.014840706251561642, + "learning_rate": 0.0006, + "loss": 4.117877960205078, + "step": 6373 + }, + { + "epoch": 88.53100436681223, + "grad_norm": 0.014165930449962616, + "learning_rate": 0.0006, + "loss": 4.1864237785339355, + "step": 6374 + }, + { + "epoch": 88.54497816593886, + "grad_norm": 0.014390201307833195, + "learning_rate": 0.0006, + "loss": 4.1044602394104, + "step": 6375 + }, + { + "epoch": 88.5589519650655, + "grad_norm": 0.015037395060062408, + "learning_rate": 0.0006, + "loss": 4.101107120513916, + "step": 6376 + }, + { + "epoch": 88.57292576419214, + "grad_norm": 0.01671452447772026, + "learning_rate": 0.0006, + "loss": 4.170581817626953, + "step": 6377 + }, + { + "epoch": 88.58689956331878, + "grad_norm": 0.01993076130747795, + "learning_rate": 0.0006, + "loss": 4.261412620544434, + "step": 6378 + }, + { + "epoch": 88.60087336244541, + "grad_norm": 0.023721016943454742, + "learning_rate": 0.0006, + "loss": 4.199363708496094, + "step": 6379 + }, + { + "epoch": 88.61484716157206, + "grad_norm": 0.023726163432002068, + "learning_rate": 0.0006, + "loss": 4.061285018920898, + "step": 6380 + }, + { + "epoch": 88.62882096069869, + "grad_norm": 0.018352877348661423, + "learning_rate": 0.0006, + "loss": 4.0721282958984375, + "step": 6381 + }, + { + "epoch": 88.64279475982532, + "grad_norm": 0.01679830253124237, + "learning_rate": 0.0006, + "loss": 4.1717939376831055, + "step": 6382 + }, + { + "epoch": 88.65676855895197, + "grad_norm": 0.017193637788295746, + "learning_rate": 0.0006, + "loss": 4.230783939361572, + "step": 6383 + }, + { + "epoch": 88.6707423580786, + "grad_norm": 0.01581464521586895, + "learning_rate": 0.0006, + "loss": 4.029906272888184, + "step": 6384 + }, + { + "epoch": 88.68471615720524, + "grad_norm": 0.014049704186618328, + "learning_rate": 0.0006, + "loss": 4.192776679992676, + "step": 6385 + }, + { + "epoch": 88.69868995633188, + "grad_norm": 0.017214450985193253, + "learning_rate": 0.0006, + "loss": 4.248916149139404, + "step": 6386 + }, + { + "epoch": 88.71266375545852, + "grad_norm": 0.0173103716224432, + "learning_rate": 0.0006, + "loss": 4.124347686767578, + "step": 6387 + }, + { + "epoch": 88.72663755458515, + "grad_norm": 0.015745092183351517, + "learning_rate": 0.0006, + "loss": 4.192878246307373, + "step": 6388 + }, + { + "epoch": 88.74061135371178, + "grad_norm": 0.013750772923231125, + "learning_rate": 0.0006, + "loss": 4.1564483642578125, + "step": 6389 + }, + { + "epoch": 88.75458515283843, + "grad_norm": 0.014969858340919018, + "learning_rate": 0.0006, + "loss": 4.0980424880981445, + "step": 6390 + }, + { + "epoch": 88.76855895196506, + "grad_norm": 0.01287173479795456, + "learning_rate": 0.0006, + "loss": 4.145712852478027, + "step": 6391 + }, + { + "epoch": 88.78253275109171, + "grad_norm": 0.012716300785541534, + "learning_rate": 0.0006, + "loss": 4.214226722717285, + "step": 6392 + }, + { + "epoch": 88.79650655021834, + "grad_norm": 0.013855204917490482, + "learning_rate": 0.0006, + "loss": 4.1667256355285645, + "step": 6393 + }, + { + "epoch": 88.81048034934499, + "grad_norm": 0.013650864362716675, + "learning_rate": 0.0006, + "loss": 4.161942481994629, + "step": 6394 + }, + { + "epoch": 88.82445414847162, + "grad_norm": 0.013780126348137856, + "learning_rate": 0.0006, + "loss": 4.087825775146484, + "step": 6395 + }, + { + "epoch": 88.83842794759825, + "grad_norm": 0.014029948972165585, + "learning_rate": 0.0006, + "loss": 4.142813205718994, + "step": 6396 + }, + { + "epoch": 88.8524017467249, + "grad_norm": 0.01322692260146141, + "learning_rate": 0.0006, + "loss": 4.106776714324951, + "step": 6397 + }, + { + "epoch": 88.86637554585153, + "grad_norm": 0.013522760011255741, + "learning_rate": 0.0006, + "loss": 4.090996265411377, + "step": 6398 + }, + { + "epoch": 88.88034934497817, + "grad_norm": 0.014474684372544289, + "learning_rate": 0.0006, + "loss": 4.124624252319336, + "step": 6399 + }, + { + "epoch": 88.8943231441048, + "grad_norm": 0.014490422792732716, + "learning_rate": 0.0006, + "loss": 4.065431594848633, + "step": 6400 + }, + { + "epoch": 88.90829694323143, + "grad_norm": 0.015379710122942924, + "learning_rate": 0.0006, + "loss": 4.207939624786377, + "step": 6401 + }, + { + "epoch": 88.92227074235808, + "grad_norm": 0.015914618968963623, + "learning_rate": 0.0006, + "loss": 4.041152000427246, + "step": 6402 + }, + { + "epoch": 88.93624454148471, + "grad_norm": 0.01480048056691885, + "learning_rate": 0.0006, + "loss": 4.138754844665527, + "step": 6403 + }, + { + "epoch": 88.95021834061136, + "grad_norm": 0.013734851032495499, + "learning_rate": 0.0006, + "loss": 4.25323486328125, + "step": 6404 + }, + { + "epoch": 88.96419213973799, + "grad_norm": 0.01452311035245657, + "learning_rate": 0.0006, + "loss": 4.115228176116943, + "step": 6405 + }, + { + "epoch": 88.97816593886463, + "grad_norm": 0.016197681427001953, + "learning_rate": 0.0006, + "loss": 4.099531173706055, + "step": 6406 + }, + { + "epoch": 88.99213973799127, + "grad_norm": 0.017169523984193802, + "learning_rate": 0.0006, + "loss": 4.182981491088867, + "step": 6407 + }, + { + "epoch": 89.0, + "grad_norm": 0.017855646088719368, + "learning_rate": 0.0006, + "loss": 4.1470136642456055, + "step": 6408 + }, + { + "epoch": 89.0, + "eval_loss": 4.6035614013671875, + "eval_runtime": 57.0625, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6408 + }, + { + "epoch": 89.01397379912663, + "grad_norm": 0.014834101311862469, + "learning_rate": 0.0006, + "loss": 4.159243583679199, + "step": 6409 + }, + { + "epoch": 89.02794759825328, + "grad_norm": 0.015316772274672985, + "learning_rate": 0.0006, + "loss": 4.0709123611450195, + "step": 6410 + }, + { + "epoch": 89.04192139737991, + "grad_norm": 0.01576591096818447, + "learning_rate": 0.0006, + "loss": 4.0921196937561035, + "step": 6411 + }, + { + "epoch": 89.05589519650655, + "grad_norm": 0.015948045998811722, + "learning_rate": 0.0006, + "loss": 4.181877136230469, + "step": 6412 + }, + { + "epoch": 89.06986899563319, + "grad_norm": 0.016612380743026733, + "learning_rate": 0.0006, + "loss": 4.081249237060547, + "step": 6413 + }, + { + "epoch": 89.08384279475983, + "grad_norm": 0.017455685883760452, + "learning_rate": 0.0006, + "loss": 4.175990104675293, + "step": 6414 + }, + { + "epoch": 89.09781659388646, + "grad_norm": 0.015040521509945393, + "learning_rate": 0.0006, + "loss": 4.121378421783447, + "step": 6415 + }, + { + "epoch": 89.1117903930131, + "grad_norm": 0.014513209462165833, + "learning_rate": 0.0006, + "loss": 3.996107816696167, + "step": 6416 + }, + { + "epoch": 89.12576419213974, + "grad_norm": 0.015455638989806175, + "learning_rate": 0.0006, + "loss": 4.073675155639648, + "step": 6417 + }, + { + "epoch": 89.13973799126637, + "grad_norm": 0.015468129888176918, + "learning_rate": 0.0006, + "loss": 4.130242824554443, + "step": 6418 + }, + { + "epoch": 89.15371179039302, + "grad_norm": 0.015284297987818718, + "learning_rate": 0.0006, + "loss": 4.18555212020874, + "step": 6419 + }, + { + "epoch": 89.16768558951965, + "grad_norm": 0.01476874016225338, + "learning_rate": 0.0006, + "loss": 4.134175777435303, + "step": 6420 + }, + { + "epoch": 89.18165938864628, + "grad_norm": 0.015836201608181, + "learning_rate": 0.0006, + "loss": 4.094529151916504, + "step": 6421 + }, + { + "epoch": 89.19563318777293, + "grad_norm": 0.0161928478628397, + "learning_rate": 0.0006, + "loss": 4.12647819519043, + "step": 6422 + }, + { + "epoch": 89.20960698689956, + "grad_norm": 0.015680814161896706, + "learning_rate": 0.0006, + "loss": 3.9620392322540283, + "step": 6423 + }, + { + "epoch": 89.2235807860262, + "grad_norm": 0.01584681309759617, + "learning_rate": 0.0006, + "loss": 4.049304008483887, + "step": 6424 + }, + { + "epoch": 89.23755458515284, + "grad_norm": 0.01452907919883728, + "learning_rate": 0.0006, + "loss": 4.064027786254883, + "step": 6425 + }, + { + "epoch": 89.25152838427948, + "grad_norm": 0.014002552255988121, + "learning_rate": 0.0006, + "loss": 4.151862621307373, + "step": 6426 + }, + { + "epoch": 89.26550218340611, + "grad_norm": 0.014658625237643719, + "learning_rate": 0.0006, + "loss": 4.1672210693359375, + "step": 6427 + }, + { + "epoch": 89.27947598253274, + "grad_norm": 0.015021766535937786, + "learning_rate": 0.0006, + "loss": 3.9742586612701416, + "step": 6428 + }, + { + "epoch": 89.29344978165939, + "grad_norm": 0.014681321568787098, + "learning_rate": 0.0006, + "loss": 4.024159908294678, + "step": 6429 + }, + { + "epoch": 89.30742358078602, + "grad_norm": 0.014161549508571625, + "learning_rate": 0.0006, + "loss": 4.105640888214111, + "step": 6430 + }, + { + "epoch": 89.32139737991267, + "grad_norm": 0.014801453799009323, + "learning_rate": 0.0006, + "loss": 4.123049736022949, + "step": 6431 + }, + { + "epoch": 89.3353711790393, + "grad_norm": 0.014307181350886822, + "learning_rate": 0.0006, + "loss": 4.086954593658447, + "step": 6432 + }, + { + "epoch": 89.34934497816595, + "grad_norm": 0.013299369253218174, + "learning_rate": 0.0006, + "loss": 4.109533786773682, + "step": 6433 + }, + { + "epoch": 89.36331877729258, + "grad_norm": 0.014388669282197952, + "learning_rate": 0.0006, + "loss": 4.222370147705078, + "step": 6434 + }, + { + "epoch": 89.37729257641921, + "grad_norm": 0.014501879923045635, + "learning_rate": 0.0006, + "loss": 4.224178314208984, + "step": 6435 + }, + { + "epoch": 89.39126637554585, + "grad_norm": 0.014109021984040737, + "learning_rate": 0.0006, + "loss": 4.112702369689941, + "step": 6436 + }, + { + "epoch": 89.40524017467249, + "grad_norm": 0.013521002605557442, + "learning_rate": 0.0006, + "loss": 4.010124683380127, + "step": 6437 + }, + { + "epoch": 89.41921397379913, + "grad_norm": 0.013288144953548908, + "learning_rate": 0.0006, + "loss": 4.127592086791992, + "step": 6438 + }, + { + "epoch": 89.43318777292576, + "grad_norm": 0.013882486149668694, + "learning_rate": 0.0006, + "loss": 4.031871318817139, + "step": 6439 + }, + { + "epoch": 89.44716157205241, + "grad_norm": 0.01289335172623396, + "learning_rate": 0.0006, + "loss": 4.0757551193237305, + "step": 6440 + }, + { + "epoch": 89.46113537117904, + "grad_norm": 0.01356302760541439, + "learning_rate": 0.0006, + "loss": 4.075678825378418, + "step": 6441 + }, + { + "epoch": 89.47510917030567, + "grad_norm": 0.01447434164583683, + "learning_rate": 0.0006, + "loss": 4.093453407287598, + "step": 6442 + }, + { + "epoch": 89.48908296943232, + "grad_norm": 0.014443481341004372, + "learning_rate": 0.0006, + "loss": 4.075372219085693, + "step": 6443 + }, + { + "epoch": 89.50305676855895, + "grad_norm": 0.01314218994230032, + "learning_rate": 0.0006, + "loss": 4.171533584594727, + "step": 6444 + }, + { + "epoch": 89.5170305676856, + "grad_norm": 0.01348346658051014, + "learning_rate": 0.0006, + "loss": 4.227236270904541, + "step": 6445 + }, + { + "epoch": 89.53100436681223, + "grad_norm": 0.013370679691433907, + "learning_rate": 0.0006, + "loss": 3.973062515258789, + "step": 6446 + }, + { + "epoch": 89.54497816593886, + "grad_norm": 0.01410642359405756, + "learning_rate": 0.0006, + "loss": 4.231175422668457, + "step": 6447 + }, + { + "epoch": 89.5589519650655, + "grad_norm": 0.01419337373226881, + "learning_rate": 0.0006, + "loss": 4.145868301391602, + "step": 6448 + }, + { + "epoch": 89.57292576419214, + "grad_norm": 0.013691053725779057, + "learning_rate": 0.0006, + "loss": 4.175131797790527, + "step": 6449 + }, + { + "epoch": 89.58689956331878, + "grad_norm": 0.013714492321014404, + "learning_rate": 0.0006, + "loss": 4.251893997192383, + "step": 6450 + }, + { + "epoch": 89.60087336244541, + "grad_norm": 0.01393197476863861, + "learning_rate": 0.0006, + "loss": 4.047079086303711, + "step": 6451 + }, + { + "epoch": 89.61484716157206, + "grad_norm": 0.013957309536635876, + "learning_rate": 0.0006, + "loss": 4.128530502319336, + "step": 6452 + }, + { + "epoch": 89.62882096069869, + "grad_norm": 0.013346714898943901, + "learning_rate": 0.0006, + "loss": 4.2296929359436035, + "step": 6453 + }, + { + "epoch": 89.64279475982532, + "grad_norm": 0.014243829064071178, + "learning_rate": 0.0006, + "loss": 4.128233909606934, + "step": 6454 + }, + { + "epoch": 89.65676855895197, + "grad_norm": 0.01606106385588646, + "learning_rate": 0.0006, + "loss": 4.202202320098877, + "step": 6455 + }, + { + "epoch": 89.6707423580786, + "grad_norm": 0.016074348241090775, + "learning_rate": 0.0006, + "loss": 4.215323448181152, + "step": 6456 + }, + { + "epoch": 89.68471615720524, + "grad_norm": 0.01611054316163063, + "learning_rate": 0.0006, + "loss": 4.096758842468262, + "step": 6457 + }, + { + "epoch": 89.69868995633188, + "grad_norm": 0.013788217678666115, + "learning_rate": 0.0006, + "loss": 4.098560810089111, + "step": 6458 + }, + { + "epoch": 89.71266375545852, + "grad_norm": 0.014408969320356846, + "learning_rate": 0.0006, + "loss": 4.107398509979248, + "step": 6459 + }, + { + "epoch": 89.72663755458515, + "grad_norm": 0.014765267260372639, + "learning_rate": 0.0006, + "loss": 4.057718276977539, + "step": 6460 + }, + { + "epoch": 89.74061135371178, + "grad_norm": 0.017395280301570892, + "learning_rate": 0.0006, + "loss": 4.100795269012451, + "step": 6461 + }, + { + "epoch": 89.75458515283843, + "grad_norm": 0.016821149736642838, + "learning_rate": 0.0006, + "loss": 4.153731346130371, + "step": 6462 + }, + { + "epoch": 89.76855895196506, + "grad_norm": 0.015633124858140945, + "learning_rate": 0.0006, + "loss": 4.061585903167725, + "step": 6463 + }, + { + "epoch": 89.78253275109171, + "grad_norm": 0.01451497245579958, + "learning_rate": 0.0006, + "loss": 4.1624884605407715, + "step": 6464 + }, + { + "epoch": 89.79650655021834, + "grad_norm": 0.013868392445147038, + "learning_rate": 0.0006, + "loss": 4.178536415100098, + "step": 6465 + }, + { + "epoch": 89.81048034934499, + "grad_norm": 0.014259875752031803, + "learning_rate": 0.0006, + "loss": 4.073114395141602, + "step": 6466 + }, + { + "epoch": 89.82445414847162, + "grad_norm": 0.013651594519615173, + "learning_rate": 0.0006, + "loss": 4.050265312194824, + "step": 6467 + }, + { + "epoch": 89.83842794759825, + "grad_norm": 0.014063458889722824, + "learning_rate": 0.0006, + "loss": 4.2014689445495605, + "step": 6468 + }, + { + "epoch": 89.8524017467249, + "grad_norm": 0.014179578050971031, + "learning_rate": 0.0006, + "loss": 4.110550880432129, + "step": 6469 + }, + { + "epoch": 89.86637554585153, + "grad_norm": 0.013616854324936867, + "learning_rate": 0.0006, + "loss": 4.1776580810546875, + "step": 6470 + }, + { + "epoch": 89.88034934497817, + "grad_norm": 0.014351330697536469, + "learning_rate": 0.0006, + "loss": 4.058526992797852, + "step": 6471 + }, + { + "epoch": 89.8943231441048, + "grad_norm": 0.0141244325786829, + "learning_rate": 0.0006, + "loss": 4.068673610687256, + "step": 6472 + }, + { + "epoch": 89.90829694323143, + "grad_norm": 0.01552131213247776, + "learning_rate": 0.0006, + "loss": 4.086787223815918, + "step": 6473 + }, + { + "epoch": 89.92227074235808, + "grad_norm": 0.014614894054830074, + "learning_rate": 0.0006, + "loss": 4.2190656661987305, + "step": 6474 + }, + { + "epoch": 89.93624454148471, + "grad_norm": 0.014369758777320385, + "learning_rate": 0.0006, + "loss": 4.117021560668945, + "step": 6475 + }, + { + "epoch": 89.95021834061136, + "grad_norm": 0.015352046117186546, + "learning_rate": 0.0006, + "loss": 4.167664527893066, + "step": 6476 + }, + { + "epoch": 89.96419213973799, + "grad_norm": 0.014314834959805012, + "learning_rate": 0.0006, + "loss": 4.160182952880859, + "step": 6477 + }, + { + "epoch": 89.97816593886463, + "grad_norm": 0.013511341996490955, + "learning_rate": 0.0006, + "loss": 4.096034049987793, + "step": 6478 + }, + { + "epoch": 89.99213973799127, + "grad_norm": 0.01308482326567173, + "learning_rate": 0.0006, + "loss": 4.062689781188965, + "step": 6479 + }, + { + "epoch": 90.0, + "grad_norm": 0.01442806702107191, + "learning_rate": 0.0006, + "loss": 4.1955885887146, + "step": 6480 + }, + { + "epoch": 90.0, + "eval_loss": 4.644347667694092, + "eval_runtime": 56.8677, + "eval_samples_per_second": 42.942, + "eval_steps_per_second": 1.354, + "step": 6480 + }, + { + "epoch": 90.01397379912663, + "grad_norm": 0.014839007519185543, + "learning_rate": 0.0006, + "loss": 4.171868324279785, + "step": 6481 + }, + { + "epoch": 90.02794759825328, + "grad_norm": 0.01476900465786457, + "learning_rate": 0.0006, + "loss": 4.119754791259766, + "step": 6482 + }, + { + "epoch": 90.04192139737991, + "grad_norm": 0.015179301612079144, + "learning_rate": 0.0006, + "loss": 4.06740140914917, + "step": 6483 + }, + { + "epoch": 90.05589519650655, + "grad_norm": 0.015552980825304985, + "learning_rate": 0.0006, + "loss": 4.112268447875977, + "step": 6484 + }, + { + "epoch": 90.06986899563319, + "grad_norm": 0.014930234290659428, + "learning_rate": 0.0006, + "loss": 4.071866989135742, + "step": 6485 + }, + { + "epoch": 90.08384279475983, + "grad_norm": 0.014124148525297642, + "learning_rate": 0.0006, + "loss": 4.013731956481934, + "step": 6486 + }, + { + "epoch": 90.09781659388646, + "grad_norm": 0.014204252511262894, + "learning_rate": 0.0006, + "loss": 4.133620738983154, + "step": 6487 + }, + { + "epoch": 90.1117903930131, + "grad_norm": 0.014845543541014194, + "learning_rate": 0.0006, + "loss": 4.122956275939941, + "step": 6488 + }, + { + "epoch": 90.12576419213974, + "grad_norm": 0.0146744754165411, + "learning_rate": 0.0006, + "loss": 4.087121963500977, + "step": 6489 + }, + { + "epoch": 90.13973799126637, + "grad_norm": 0.016007190570235252, + "learning_rate": 0.0006, + "loss": 3.98740816116333, + "step": 6490 + }, + { + "epoch": 90.15371179039302, + "grad_norm": 0.018072661012411118, + "learning_rate": 0.0006, + "loss": 4.1155009269714355, + "step": 6491 + }, + { + "epoch": 90.16768558951965, + "grad_norm": 0.016868166625499725, + "learning_rate": 0.0006, + "loss": 4.123727321624756, + "step": 6492 + }, + { + "epoch": 90.18165938864628, + "grad_norm": 0.017091726884245872, + "learning_rate": 0.0006, + "loss": 4.1405229568481445, + "step": 6493 + }, + { + "epoch": 90.19563318777293, + "grad_norm": 0.018329855054616928, + "learning_rate": 0.0006, + "loss": 4.093900680541992, + "step": 6494 + }, + { + "epoch": 90.20960698689956, + "grad_norm": 0.018906457349658012, + "learning_rate": 0.0006, + "loss": 4.0552849769592285, + "step": 6495 + }, + { + "epoch": 90.2235807860262, + "grad_norm": 0.017275869846343994, + "learning_rate": 0.0006, + "loss": 4.1310625076293945, + "step": 6496 + }, + { + "epoch": 90.23755458515284, + "grad_norm": 0.015029897913336754, + "learning_rate": 0.0006, + "loss": 4.100166320800781, + "step": 6497 + }, + { + "epoch": 90.25152838427948, + "grad_norm": 0.014891199767589569, + "learning_rate": 0.0006, + "loss": 4.1314802169799805, + "step": 6498 + }, + { + "epoch": 90.26550218340611, + "grad_norm": 0.014877690002322197, + "learning_rate": 0.0006, + "loss": 4.076098442077637, + "step": 6499 + }, + { + "epoch": 90.27947598253274, + "grad_norm": 0.014670864678919315, + "learning_rate": 0.0006, + "loss": 4.066137313842773, + "step": 6500 + }, + { + "epoch": 90.29344978165939, + "grad_norm": 0.01444234699010849, + "learning_rate": 0.0006, + "loss": 4.013591766357422, + "step": 6501 + }, + { + "epoch": 90.30742358078602, + "grad_norm": 0.01339263841509819, + "learning_rate": 0.0006, + "loss": 4.117959499359131, + "step": 6502 + }, + { + "epoch": 90.32139737991267, + "grad_norm": 0.014753810130059719, + "learning_rate": 0.0006, + "loss": 4.042947769165039, + "step": 6503 + }, + { + "epoch": 90.3353711790393, + "grad_norm": 0.014947964809834957, + "learning_rate": 0.0006, + "loss": 4.112810134887695, + "step": 6504 + }, + { + "epoch": 90.34934497816595, + "grad_norm": 0.015155353583395481, + "learning_rate": 0.0006, + "loss": 4.213214874267578, + "step": 6505 + }, + { + "epoch": 90.36331877729258, + "grad_norm": 0.015502216294407845, + "learning_rate": 0.0006, + "loss": 4.127619743347168, + "step": 6506 + }, + { + "epoch": 90.37729257641921, + "grad_norm": 0.015377781353890896, + "learning_rate": 0.0006, + "loss": 4.135862350463867, + "step": 6507 + }, + { + "epoch": 90.39126637554585, + "grad_norm": 0.01662006787955761, + "learning_rate": 0.0006, + "loss": 4.076877593994141, + "step": 6508 + }, + { + "epoch": 90.40524017467249, + "grad_norm": 0.016324395313858986, + "learning_rate": 0.0006, + "loss": 4.214895725250244, + "step": 6509 + }, + { + "epoch": 90.41921397379913, + "grad_norm": 0.016099024564027786, + "learning_rate": 0.0006, + "loss": 4.139717102050781, + "step": 6510 + }, + { + "epoch": 90.43318777292576, + "grad_norm": 0.014403503388166428, + "learning_rate": 0.0006, + "loss": 4.040343284606934, + "step": 6511 + }, + { + "epoch": 90.44716157205241, + "grad_norm": 0.01413416862487793, + "learning_rate": 0.0006, + "loss": 4.087702751159668, + "step": 6512 + }, + { + "epoch": 90.46113537117904, + "grad_norm": 0.01603989489376545, + "learning_rate": 0.0006, + "loss": 4.173558712005615, + "step": 6513 + }, + { + "epoch": 90.47510917030567, + "grad_norm": 0.016846206039190292, + "learning_rate": 0.0006, + "loss": 4.1955461502075195, + "step": 6514 + }, + { + "epoch": 90.48908296943232, + "grad_norm": 0.01652812585234642, + "learning_rate": 0.0006, + "loss": 4.061838150024414, + "step": 6515 + }, + { + "epoch": 90.50305676855895, + "grad_norm": 0.015100528486073017, + "learning_rate": 0.0006, + "loss": 4.041808605194092, + "step": 6516 + }, + { + "epoch": 90.5170305676856, + "grad_norm": 0.013867822475731373, + "learning_rate": 0.0006, + "loss": 4.12662410736084, + "step": 6517 + }, + { + "epoch": 90.53100436681223, + "grad_norm": 0.013684107922017574, + "learning_rate": 0.0006, + "loss": 4.1254048347473145, + "step": 6518 + }, + { + "epoch": 90.54497816593886, + "grad_norm": 0.014200013130903244, + "learning_rate": 0.0006, + "loss": 4.183216094970703, + "step": 6519 + }, + { + "epoch": 90.5589519650655, + "grad_norm": 0.013707022182643414, + "learning_rate": 0.0006, + "loss": 4.07176399230957, + "step": 6520 + }, + { + "epoch": 90.57292576419214, + "grad_norm": 0.013996582478284836, + "learning_rate": 0.0006, + "loss": 4.006575584411621, + "step": 6521 + }, + { + "epoch": 90.58689956331878, + "grad_norm": 0.014886210672557354, + "learning_rate": 0.0006, + "loss": 4.118956565856934, + "step": 6522 + }, + { + "epoch": 90.60087336244541, + "grad_norm": 0.014022685587406158, + "learning_rate": 0.0006, + "loss": 3.96986722946167, + "step": 6523 + }, + { + "epoch": 90.61484716157206, + "grad_norm": 0.014547634869813919, + "learning_rate": 0.0006, + "loss": 4.122886657714844, + "step": 6524 + }, + { + "epoch": 90.62882096069869, + "grad_norm": 0.013682851567864418, + "learning_rate": 0.0006, + "loss": 3.988292694091797, + "step": 6525 + }, + { + "epoch": 90.64279475982532, + "grad_norm": 0.014016421511769295, + "learning_rate": 0.0006, + "loss": 4.110270977020264, + "step": 6526 + }, + { + "epoch": 90.65676855895197, + "grad_norm": 0.013392233289778233, + "learning_rate": 0.0006, + "loss": 4.065059661865234, + "step": 6527 + }, + { + "epoch": 90.6707423580786, + "grad_norm": 0.012107795104384422, + "learning_rate": 0.0006, + "loss": 4.034260272979736, + "step": 6528 + }, + { + "epoch": 90.68471615720524, + "grad_norm": 0.013198604807257652, + "learning_rate": 0.0006, + "loss": 4.162083625793457, + "step": 6529 + }, + { + "epoch": 90.69868995633188, + "grad_norm": 0.012380579486489296, + "learning_rate": 0.0006, + "loss": 4.040456295013428, + "step": 6530 + }, + { + "epoch": 90.71266375545852, + "grad_norm": 0.012784135527908802, + "learning_rate": 0.0006, + "loss": 4.0823798179626465, + "step": 6531 + }, + { + "epoch": 90.72663755458515, + "grad_norm": 0.012920220382511616, + "learning_rate": 0.0006, + "loss": 4.181064128875732, + "step": 6532 + }, + { + "epoch": 90.74061135371178, + "grad_norm": 0.01338814664632082, + "learning_rate": 0.0006, + "loss": 4.111660003662109, + "step": 6533 + }, + { + "epoch": 90.75458515283843, + "grad_norm": 0.013628040440380573, + "learning_rate": 0.0006, + "loss": 4.237852573394775, + "step": 6534 + }, + { + "epoch": 90.76855895196506, + "grad_norm": 0.013384854421019554, + "learning_rate": 0.0006, + "loss": 4.121862888336182, + "step": 6535 + }, + { + "epoch": 90.78253275109171, + "grad_norm": 0.014726929366588593, + "learning_rate": 0.0006, + "loss": 4.1695990562438965, + "step": 6536 + }, + { + "epoch": 90.79650655021834, + "grad_norm": 0.0161006897687912, + "learning_rate": 0.0006, + "loss": 4.17794132232666, + "step": 6537 + }, + { + "epoch": 90.81048034934499, + "grad_norm": 0.016761884093284607, + "learning_rate": 0.0006, + "loss": 4.093764781951904, + "step": 6538 + }, + { + "epoch": 90.82445414847162, + "grad_norm": 0.017246825620532036, + "learning_rate": 0.0006, + "loss": 3.9382331371307373, + "step": 6539 + }, + { + "epoch": 90.83842794759825, + "grad_norm": 0.018141290172934532, + "learning_rate": 0.0006, + "loss": 3.973609209060669, + "step": 6540 + }, + { + "epoch": 90.8524017467249, + "grad_norm": 0.018129250034689903, + "learning_rate": 0.0006, + "loss": 4.227414131164551, + "step": 6541 + }, + { + "epoch": 90.86637554585153, + "grad_norm": 0.018234573304653168, + "learning_rate": 0.0006, + "loss": 4.151870250701904, + "step": 6542 + }, + { + "epoch": 90.88034934497817, + "grad_norm": 0.01560172438621521, + "learning_rate": 0.0006, + "loss": 4.127397537231445, + "step": 6543 + }, + { + "epoch": 90.8943231441048, + "grad_norm": 0.014367681927978992, + "learning_rate": 0.0006, + "loss": 4.17820930480957, + "step": 6544 + }, + { + "epoch": 90.90829694323143, + "grad_norm": 0.016071319580078125, + "learning_rate": 0.0006, + "loss": 4.059854507446289, + "step": 6545 + }, + { + "epoch": 90.92227074235808, + "grad_norm": 0.01589169353246689, + "learning_rate": 0.0006, + "loss": 4.282748222351074, + "step": 6546 + }, + { + "epoch": 90.93624454148471, + "grad_norm": 0.013639772310853004, + "learning_rate": 0.0006, + "loss": 4.147368907928467, + "step": 6547 + }, + { + "epoch": 90.95021834061136, + "grad_norm": 0.014248427003622055, + "learning_rate": 0.0006, + "loss": 4.240248680114746, + "step": 6548 + }, + { + "epoch": 90.96419213973799, + "grad_norm": 0.014600320719182491, + "learning_rate": 0.0006, + "loss": 4.122123718261719, + "step": 6549 + }, + { + "epoch": 90.97816593886463, + "grad_norm": 0.014371675439178944, + "learning_rate": 0.0006, + "loss": 4.013302803039551, + "step": 6550 + }, + { + "epoch": 90.99213973799127, + "grad_norm": 0.013133526779711246, + "learning_rate": 0.0006, + "loss": 4.097136497497559, + "step": 6551 + }, + { + "epoch": 91.0, + "grad_norm": 0.015599353238940239, + "learning_rate": 0.0006, + "loss": 4.198554039001465, + "step": 6552 + }, + { + "epoch": 91.0, + "eval_loss": 4.603907585144043, + "eval_runtime": 56.9418, + "eval_samples_per_second": 42.886, + "eval_steps_per_second": 1.352, + "step": 6552 + }, + { + "epoch": 91.01397379912663, + "grad_norm": 0.015917908400297165, + "learning_rate": 0.0006, + "loss": 4.133877754211426, + "step": 6553 + }, + { + "epoch": 91.02794759825328, + "grad_norm": 0.015485800802707672, + "learning_rate": 0.0006, + "loss": 4.111617088317871, + "step": 6554 + }, + { + "epoch": 91.04192139737991, + "grad_norm": 0.012969603762030602, + "learning_rate": 0.0006, + "loss": 4.064828395843506, + "step": 6555 + }, + { + "epoch": 91.05589519650655, + "grad_norm": 0.014482097700238228, + "learning_rate": 0.0006, + "loss": 4.101710319519043, + "step": 6556 + }, + { + "epoch": 91.06986899563319, + "grad_norm": 0.017242740839719772, + "learning_rate": 0.0006, + "loss": 4.015342712402344, + "step": 6557 + }, + { + "epoch": 91.08384279475983, + "grad_norm": 0.017100440338253975, + "learning_rate": 0.0006, + "loss": 3.9801602363586426, + "step": 6558 + }, + { + "epoch": 91.09781659388646, + "grad_norm": 0.015969131141901016, + "learning_rate": 0.0006, + "loss": 4.049958229064941, + "step": 6559 + }, + { + "epoch": 91.1117903930131, + "grad_norm": 0.016018927097320557, + "learning_rate": 0.0006, + "loss": 4.094139099121094, + "step": 6560 + }, + { + "epoch": 91.12576419213974, + "grad_norm": 0.018131926655769348, + "learning_rate": 0.0006, + "loss": 3.9499974250793457, + "step": 6561 + }, + { + "epoch": 91.13973799126637, + "grad_norm": 0.019425220787525177, + "learning_rate": 0.0006, + "loss": 3.9910552501678467, + "step": 6562 + }, + { + "epoch": 91.15371179039302, + "grad_norm": 0.01908663846552372, + "learning_rate": 0.0006, + "loss": 4.032179832458496, + "step": 6563 + }, + { + "epoch": 91.16768558951965, + "grad_norm": 0.01871207356452942, + "learning_rate": 0.0006, + "loss": 4.112380027770996, + "step": 6564 + }, + { + "epoch": 91.18165938864628, + "grad_norm": 0.020867953076958656, + "learning_rate": 0.0006, + "loss": 4.06105899810791, + "step": 6565 + }, + { + "epoch": 91.19563318777293, + "grad_norm": 0.019053412601351738, + "learning_rate": 0.0006, + "loss": 3.992504596710205, + "step": 6566 + }, + { + "epoch": 91.20960698689956, + "grad_norm": 0.016084402799606323, + "learning_rate": 0.0006, + "loss": 4.100559234619141, + "step": 6567 + }, + { + "epoch": 91.2235807860262, + "grad_norm": 0.016857590526342392, + "learning_rate": 0.0006, + "loss": 4.1567277908325195, + "step": 6568 + }, + { + "epoch": 91.23755458515284, + "grad_norm": 0.01702957972884178, + "learning_rate": 0.0006, + "loss": 4.234771728515625, + "step": 6569 + }, + { + "epoch": 91.25152838427948, + "grad_norm": 0.016869494691491127, + "learning_rate": 0.0006, + "loss": 3.9931387901306152, + "step": 6570 + }, + { + "epoch": 91.26550218340611, + "grad_norm": 0.01696581020951271, + "learning_rate": 0.0006, + "loss": 4.184908866882324, + "step": 6571 + }, + { + "epoch": 91.27947598253274, + "grad_norm": 0.014982126653194427, + "learning_rate": 0.0006, + "loss": 4.036668300628662, + "step": 6572 + }, + { + "epoch": 91.29344978165939, + "grad_norm": 0.014363301917910576, + "learning_rate": 0.0006, + "loss": 4.074010848999023, + "step": 6573 + }, + { + "epoch": 91.30742358078602, + "grad_norm": 0.015492337755858898, + "learning_rate": 0.0006, + "loss": 4.127626419067383, + "step": 6574 + }, + { + "epoch": 91.32139737991267, + "grad_norm": 0.015375050716102123, + "learning_rate": 0.0006, + "loss": 4.159229755401611, + "step": 6575 + }, + { + "epoch": 91.3353711790393, + "grad_norm": 0.015068565495312214, + "learning_rate": 0.0006, + "loss": 4.172623634338379, + "step": 6576 + }, + { + "epoch": 91.34934497816595, + "grad_norm": 0.0140256742015481, + "learning_rate": 0.0006, + "loss": 4.207065105438232, + "step": 6577 + }, + { + "epoch": 91.36331877729258, + "grad_norm": 0.014220152981579304, + "learning_rate": 0.0006, + "loss": 4.131444931030273, + "step": 6578 + }, + { + "epoch": 91.37729257641921, + "grad_norm": 0.0143149234354496, + "learning_rate": 0.0006, + "loss": 4.065218448638916, + "step": 6579 + }, + { + "epoch": 91.39126637554585, + "grad_norm": 0.013149659149348736, + "learning_rate": 0.0006, + "loss": 4.106743335723877, + "step": 6580 + }, + { + "epoch": 91.40524017467249, + "grad_norm": 0.014550330117344856, + "learning_rate": 0.0006, + "loss": 4.123722553253174, + "step": 6581 + }, + { + "epoch": 91.41921397379913, + "grad_norm": 0.014611710794270039, + "learning_rate": 0.0006, + "loss": 4.079610824584961, + "step": 6582 + }, + { + "epoch": 91.43318777292576, + "grad_norm": 0.013369769789278507, + "learning_rate": 0.0006, + "loss": 4.133187770843506, + "step": 6583 + }, + { + "epoch": 91.44716157205241, + "grad_norm": 0.013243545778095722, + "learning_rate": 0.0006, + "loss": 4.042607307434082, + "step": 6584 + }, + { + "epoch": 91.46113537117904, + "grad_norm": 0.013388266786932945, + "learning_rate": 0.0006, + "loss": 4.14863395690918, + "step": 6585 + }, + { + "epoch": 91.47510917030567, + "grad_norm": 0.014333273284137249, + "learning_rate": 0.0006, + "loss": 4.109699726104736, + "step": 6586 + }, + { + "epoch": 91.48908296943232, + "grad_norm": 0.012742757797241211, + "learning_rate": 0.0006, + "loss": 4.126582145690918, + "step": 6587 + }, + { + "epoch": 91.50305676855895, + "grad_norm": 0.012880304828286171, + "learning_rate": 0.0006, + "loss": 4.205300331115723, + "step": 6588 + }, + { + "epoch": 91.5170305676856, + "grad_norm": 0.013896043412387371, + "learning_rate": 0.0006, + "loss": 4.132510185241699, + "step": 6589 + }, + { + "epoch": 91.53100436681223, + "grad_norm": 0.014361281879246235, + "learning_rate": 0.0006, + "loss": 4.1487932205200195, + "step": 6590 + }, + { + "epoch": 91.54497816593886, + "grad_norm": 0.015137741342186928, + "learning_rate": 0.0006, + "loss": 4.1564531326293945, + "step": 6591 + }, + { + "epoch": 91.5589519650655, + "grad_norm": 0.014660682529211044, + "learning_rate": 0.0006, + "loss": 4.056064128875732, + "step": 6592 + }, + { + "epoch": 91.57292576419214, + "grad_norm": 0.014888476580381393, + "learning_rate": 0.0006, + "loss": 4.0633344650268555, + "step": 6593 + }, + { + "epoch": 91.58689956331878, + "grad_norm": 0.012162433005869389, + "learning_rate": 0.0006, + "loss": 4.026293754577637, + "step": 6594 + }, + { + "epoch": 91.60087336244541, + "grad_norm": 0.012892271392047405, + "learning_rate": 0.0006, + "loss": 4.267086029052734, + "step": 6595 + }, + { + "epoch": 91.61484716157206, + "grad_norm": 0.01450759544968605, + "learning_rate": 0.0006, + "loss": 4.149359703063965, + "step": 6596 + }, + { + "epoch": 91.62882096069869, + "grad_norm": 0.01419843453913927, + "learning_rate": 0.0006, + "loss": 4.094183921813965, + "step": 6597 + }, + { + "epoch": 91.64279475982532, + "grad_norm": 0.014401616528630257, + "learning_rate": 0.0006, + "loss": 4.125172138214111, + "step": 6598 + }, + { + "epoch": 91.65676855895197, + "grad_norm": 0.013945513404905796, + "learning_rate": 0.0006, + "loss": 4.146623134613037, + "step": 6599 + }, + { + "epoch": 91.6707423580786, + "grad_norm": 0.013516047969460487, + "learning_rate": 0.0006, + "loss": 4.160526275634766, + "step": 6600 + }, + { + "epoch": 91.68471615720524, + "grad_norm": 0.013262378051877022, + "learning_rate": 0.0006, + "loss": 4.074923038482666, + "step": 6601 + }, + { + "epoch": 91.69868995633188, + "grad_norm": 0.014365669339895248, + "learning_rate": 0.0006, + "loss": 4.118094444274902, + "step": 6602 + }, + { + "epoch": 91.71266375545852, + "grad_norm": 0.016713734716176987, + "learning_rate": 0.0006, + "loss": 4.021799087524414, + "step": 6603 + }, + { + "epoch": 91.72663755458515, + "grad_norm": 0.01988680474460125, + "learning_rate": 0.0006, + "loss": 4.238436222076416, + "step": 6604 + }, + { + "epoch": 91.74061135371178, + "grad_norm": 0.022749733179807663, + "learning_rate": 0.0006, + "loss": 4.010470390319824, + "step": 6605 + }, + { + "epoch": 91.75458515283843, + "grad_norm": 0.022533461451530457, + "learning_rate": 0.0006, + "loss": 4.152201175689697, + "step": 6606 + }, + { + "epoch": 91.76855895196506, + "grad_norm": 0.02021082676947117, + "learning_rate": 0.0006, + "loss": 4.120572090148926, + "step": 6607 + }, + { + "epoch": 91.78253275109171, + "grad_norm": 0.016298197209835052, + "learning_rate": 0.0006, + "loss": 4.115815162658691, + "step": 6608 + }, + { + "epoch": 91.79650655021834, + "grad_norm": 0.017562786117196083, + "learning_rate": 0.0006, + "loss": 4.109724044799805, + "step": 6609 + }, + { + "epoch": 91.81048034934499, + "grad_norm": 0.018153153359889984, + "learning_rate": 0.0006, + "loss": 4.143759727478027, + "step": 6610 + }, + { + "epoch": 91.82445414847162, + "grad_norm": 0.014762306585907936, + "learning_rate": 0.0006, + "loss": 4.25596809387207, + "step": 6611 + }, + { + "epoch": 91.83842794759825, + "grad_norm": 0.015074343420565128, + "learning_rate": 0.0006, + "loss": 4.207723140716553, + "step": 6612 + }, + { + "epoch": 91.8524017467249, + "grad_norm": 0.015028579160571098, + "learning_rate": 0.0006, + "loss": 4.01950216293335, + "step": 6613 + }, + { + "epoch": 91.86637554585153, + "grad_norm": 0.013996795751154423, + "learning_rate": 0.0006, + "loss": 4.132997989654541, + "step": 6614 + }, + { + "epoch": 91.88034934497817, + "grad_norm": 0.014978236518800259, + "learning_rate": 0.0006, + "loss": 4.132402420043945, + "step": 6615 + }, + { + "epoch": 91.8943231441048, + "grad_norm": 0.013111517764627934, + "learning_rate": 0.0006, + "loss": 4.1223649978637695, + "step": 6616 + }, + { + "epoch": 91.90829694323143, + "grad_norm": 0.013872234150767326, + "learning_rate": 0.0006, + "loss": 4.096590042114258, + "step": 6617 + }, + { + "epoch": 91.92227074235808, + "grad_norm": 0.013642470352351665, + "learning_rate": 0.0006, + "loss": 4.236227035522461, + "step": 6618 + }, + { + "epoch": 91.93624454148471, + "grad_norm": 0.014684873633086681, + "learning_rate": 0.0006, + "loss": 4.037961959838867, + "step": 6619 + }, + { + "epoch": 91.95021834061136, + "grad_norm": 0.014498206786811352, + "learning_rate": 0.0006, + "loss": 4.043325424194336, + "step": 6620 + }, + { + "epoch": 91.96419213973799, + "grad_norm": 0.012768656946718693, + "learning_rate": 0.0006, + "loss": 4.138803482055664, + "step": 6621 + }, + { + "epoch": 91.97816593886463, + "grad_norm": 0.012155905365943909, + "learning_rate": 0.0006, + "loss": 4.163424968719482, + "step": 6622 + }, + { + "epoch": 91.99213973799127, + "grad_norm": 0.012584930285811424, + "learning_rate": 0.0006, + "loss": 4.086629867553711, + "step": 6623 + }, + { + "epoch": 92.0, + "grad_norm": 0.01578151248395443, + "learning_rate": 0.0006, + "loss": 4.108386993408203, + "step": 6624 + }, + { + "epoch": 92.0, + "eval_loss": 4.541166305541992, + "eval_runtime": 56.871, + "eval_samples_per_second": 42.939, + "eval_steps_per_second": 1.354, + "step": 6624 + }, + { + "epoch": 92.01397379912663, + "grad_norm": 0.016901545226573944, + "learning_rate": 0.0006, + "loss": 4.043694496154785, + "step": 6625 + }, + { + "epoch": 92.02794759825328, + "grad_norm": 0.016299039125442505, + "learning_rate": 0.0006, + "loss": 4.100504398345947, + "step": 6626 + }, + { + "epoch": 92.04192139737991, + "grad_norm": 0.013988777063786983, + "learning_rate": 0.0006, + "loss": 4.063943862915039, + "step": 6627 + }, + { + "epoch": 92.05589519650655, + "grad_norm": 0.01343096699565649, + "learning_rate": 0.0006, + "loss": 4.0321197509765625, + "step": 6628 + }, + { + "epoch": 92.06986899563319, + "grad_norm": 0.014343147166073322, + "learning_rate": 0.0006, + "loss": 4.095499038696289, + "step": 6629 + }, + { + "epoch": 92.08384279475983, + "grad_norm": 0.014989756047725677, + "learning_rate": 0.0006, + "loss": 4.122310638427734, + "step": 6630 + }, + { + "epoch": 92.09781659388646, + "grad_norm": 0.013929174281656742, + "learning_rate": 0.0006, + "loss": 4.02870512008667, + "step": 6631 + }, + { + "epoch": 92.1117903930131, + "grad_norm": 0.014139998704195023, + "learning_rate": 0.0006, + "loss": 4.180675506591797, + "step": 6632 + }, + { + "epoch": 92.12576419213974, + "grad_norm": 0.012920474633574486, + "learning_rate": 0.0006, + "loss": 4.111686706542969, + "step": 6633 + }, + { + "epoch": 92.13973799126637, + "grad_norm": 0.01291937381029129, + "learning_rate": 0.0006, + "loss": 3.958092212677002, + "step": 6634 + }, + { + "epoch": 92.15371179039302, + "grad_norm": 0.013250474818050861, + "learning_rate": 0.0006, + "loss": 4.075105667114258, + "step": 6635 + }, + { + "epoch": 92.16768558951965, + "grad_norm": 0.01321406476199627, + "learning_rate": 0.0006, + "loss": 4.072322368621826, + "step": 6636 + }, + { + "epoch": 92.18165938864628, + "grad_norm": 0.014002865180373192, + "learning_rate": 0.0006, + "loss": 4.061572074890137, + "step": 6637 + }, + { + "epoch": 92.19563318777293, + "grad_norm": 0.013876105658710003, + "learning_rate": 0.0006, + "loss": 4.099213600158691, + "step": 6638 + }, + { + "epoch": 92.20960698689956, + "grad_norm": 0.01464459951967001, + "learning_rate": 0.0006, + "loss": 4.1573638916015625, + "step": 6639 + }, + { + "epoch": 92.2235807860262, + "grad_norm": 0.015354706905782223, + "learning_rate": 0.0006, + "loss": 4.078714370727539, + "step": 6640 + }, + { + "epoch": 92.23755458515284, + "grad_norm": 0.015265689231455326, + "learning_rate": 0.0006, + "loss": 4.036652565002441, + "step": 6641 + }, + { + "epoch": 92.25152838427948, + "grad_norm": 0.016140855848789215, + "learning_rate": 0.0006, + "loss": 4.120541095733643, + "step": 6642 + }, + { + "epoch": 92.26550218340611, + "grad_norm": 0.017044052481651306, + "learning_rate": 0.0006, + "loss": 4.015849590301514, + "step": 6643 + }, + { + "epoch": 92.27947598253274, + "grad_norm": 0.01469327136874199, + "learning_rate": 0.0006, + "loss": 4.137286186218262, + "step": 6644 + }, + { + "epoch": 92.29344978165939, + "grad_norm": 0.014596942812204361, + "learning_rate": 0.0006, + "loss": 4.11309814453125, + "step": 6645 + }, + { + "epoch": 92.30742358078602, + "grad_norm": 0.014671559445559978, + "learning_rate": 0.0006, + "loss": 4.087950229644775, + "step": 6646 + }, + { + "epoch": 92.32139737991267, + "grad_norm": 0.015040507540106773, + "learning_rate": 0.0006, + "loss": 3.967053174972534, + "step": 6647 + }, + { + "epoch": 92.3353711790393, + "grad_norm": 0.01417640782892704, + "learning_rate": 0.0006, + "loss": 3.990346908569336, + "step": 6648 + }, + { + "epoch": 92.34934497816595, + "grad_norm": 0.013855187222361565, + "learning_rate": 0.0006, + "loss": 3.965132474899292, + "step": 6649 + }, + { + "epoch": 92.36331877729258, + "grad_norm": 0.014291216619312763, + "learning_rate": 0.0006, + "loss": 4.161717891693115, + "step": 6650 + }, + { + "epoch": 92.37729257641921, + "grad_norm": 0.014148853719234467, + "learning_rate": 0.0006, + "loss": 4.019951820373535, + "step": 6651 + }, + { + "epoch": 92.39126637554585, + "grad_norm": 0.01414303295314312, + "learning_rate": 0.0006, + "loss": 4.125976085662842, + "step": 6652 + }, + { + "epoch": 92.40524017467249, + "grad_norm": 0.01358176488429308, + "learning_rate": 0.0006, + "loss": 4.088022232055664, + "step": 6653 + }, + { + "epoch": 92.41921397379913, + "grad_norm": 0.014174354262650013, + "learning_rate": 0.0006, + "loss": 4.1137518882751465, + "step": 6654 + }, + { + "epoch": 92.43318777292576, + "grad_norm": 0.013756301254034042, + "learning_rate": 0.0006, + "loss": 4.0088725090026855, + "step": 6655 + }, + { + "epoch": 92.44716157205241, + "grad_norm": 0.014257770963013172, + "learning_rate": 0.0006, + "loss": 4.05248498916626, + "step": 6656 + }, + { + "epoch": 92.46113537117904, + "grad_norm": 0.015319934114813805, + "learning_rate": 0.0006, + "loss": 4.1375732421875, + "step": 6657 + }, + { + "epoch": 92.47510917030567, + "grad_norm": 0.014655457809567451, + "learning_rate": 0.0006, + "loss": 4.172857284545898, + "step": 6658 + }, + { + "epoch": 92.48908296943232, + "grad_norm": 0.014981928281486034, + "learning_rate": 0.0006, + "loss": 4.064846038818359, + "step": 6659 + }, + { + "epoch": 92.50305676855895, + "grad_norm": 0.015197387896478176, + "learning_rate": 0.0006, + "loss": 4.119397163391113, + "step": 6660 + }, + { + "epoch": 92.5170305676856, + "grad_norm": 0.016060935333371162, + "learning_rate": 0.0006, + "loss": 4.09212064743042, + "step": 6661 + }, + { + "epoch": 92.53100436681223, + "grad_norm": 0.014884199015796185, + "learning_rate": 0.0006, + "loss": 4.027483940124512, + "step": 6662 + }, + { + "epoch": 92.54497816593886, + "grad_norm": 0.01562194712460041, + "learning_rate": 0.0006, + "loss": 3.9722909927368164, + "step": 6663 + }, + { + "epoch": 92.5589519650655, + "grad_norm": 0.014801043085753918, + "learning_rate": 0.0006, + "loss": 4.048130989074707, + "step": 6664 + }, + { + "epoch": 92.57292576419214, + "grad_norm": 0.013954494148492813, + "learning_rate": 0.0006, + "loss": 4.050351619720459, + "step": 6665 + }, + { + "epoch": 92.58689956331878, + "grad_norm": 0.013292429968714714, + "learning_rate": 0.0006, + "loss": 4.245860576629639, + "step": 6666 + }, + { + "epoch": 92.60087336244541, + "grad_norm": 0.015737095847725868, + "learning_rate": 0.0006, + "loss": 4.154323101043701, + "step": 6667 + }, + { + "epoch": 92.61484716157206, + "grad_norm": 0.018242565914988518, + "learning_rate": 0.0006, + "loss": 4.118710517883301, + "step": 6668 + }, + { + "epoch": 92.62882096069869, + "grad_norm": 0.017666636034846306, + "learning_rate": 0.0006, + "loss": 4.1013875007629395, + "step": 6669 + }, + { + "epoch": 92.64279475982532, + "grad_norm": 0.015017223544418812, + "learning_rate": 0.0006, + "loss": 4.084001064300537, + "step": 6670 + }, + { + "epoch": 92.65676855895197, + "grad_norm": 0.01359238475561142, + "learning_rate": 0.0006, + "loss": 4.091456413269043, + "step": 6671 + }, + { + "epoch": 92.6707423580786, + "grad_norm": 0.014524881727993488, + "learning_rate": 0.0006, + "loss": 4.121245384216309, + "step": 6672 + }, + { + "epoch": 92.68471615720524, + "grad_norm": 0.016214480623602867, + "learning_rate": 0.0006, + "loss": 4.133184432983398, + "step": 6673 + }, + { + "epoch": 92.69868995633188, + "grad_norm": 0.0176643505692482, + "learning_rate": 0.0006, + "loss": 4.142827987670898, + "step": 6674 + }, + { + "epoch": 92.71266375545852, + "grad_norm": 0.017004093155264854, + "learning_rate": 0.0006, + "loss": 4.067666530609131, + "step": 6675 + }, + { + "epoch": 92.72663755458515, + "grad_norm": 0.014077397994697094, + "learning_rate": 0.0006, + "loss": 4.076465129852295, + "step": 6676 + }, + { + "epoch": 92.74061135371178, + "grad_norm": 0.015905627980828285, + "learning_rate": 0.0006, + "loss": 4.083726406097412, + "step": 6677 + }, + { + "epoch": 92.75458515283843, + "grad_norm": 0.016467567533254623, + "learning_rate": 0.0006, + "loss": 4.241217136383057, + "step": 6678 + }, + { + "epoch": 92.76855895196506, + "grad_norm": 0.01541806198656559, + "learning_rate": 0.0006, + "loss": 4.05365514755249, + "step": 6679 + }, + { + "epoch": 92.78253275109171, + "grad_norm": 0.015028242021799088, + "learning_rate": 0.0006, + "loss": 4.029950141906738, + "step": 6680 + }, + { + "epoch": 92.79650655021834, + "grad_norm": 0.014366233721375465, + "learning_rate": 0.0006, + "loss": 4.181772232055664, + "step": 6681 + }, + { + "epoch": 92.81048034934499, + "grad_norm": 0.014313179068267345, + "learning_rate": 0.0006, + "loss": 4.152658462524414, + "step": 6682 + }, + { + "epoch": 92.82445414847162, + "grad_norm": 0.016504302620887756, + "learning_rate": 0.0006, + "loss": 4.080532073974609, + "step": 6683 + }, + { + "epoch": 92.83842794759825, + "grad_norm": 0.01685677096247673, + "learning_rate": 0.0006, + "loss": 4.194268226623535, + "step": 6684 + }, + { + "epoch": 92.8524017467249, + "grad_norm": 0.018684150651097298, + "learning_rate": 0.0006, + "loss": 4.218497276306152, + "step": 6685 + }, + { + "epoch": 92.86637554585153, + "grad_norm": 0.01880732551217079, + "learning_rate": 0.0006, + "loss": 4.133942604064941, + "step": 6686 + }, + { + "epoch": 92.88034934497817, + "grad_norm": 0.016504917293787003, + "learning_rate": 0.0006, + "loss": 4.139945030212402, + "step": 6687 + }, + { + "epoch": 92.8943231441048, + "grad_norm": 0.016495356336236, + "learning_rate": 0.0006, + "loss": 4.098918914794922, + "step": 6688 + }, + { + "epoch": 92.90829694323143, + "grad_norm": 0.0170587245374918, + "learning_rate": 0.0006, + "loss": 4.025818824768066, + "step": 6689 + }, + { + "epoch": 92.92227074235808, + "grad_norm": 0.01718820631504059, + "learning_rate": 0.0006, + "loss": 4.125217437744141, + "step": 6690 + }, + { + "epoch": 92.93624454148471, + "grad_norm": 0.015789354220032692, + "learning_rate": 0.0006, + "loss": 4.150068283081055, + "step": 6691 + }, + { + "epoch": 92.95021834061136, + "grad_norm": 0.01587480492889881, + "learning_rate": 0.0006, + "loss": 4.194794654846191, + "step": 6692 + }, + { + "epoch": 92.96419213973799, + "grad_norm": 0.015669209882616997, + "learning_rate": 0.0006, + "loss": 4.0550336837768555, + "step": 6693 + }, + { + "epoch": 92.97816593886463, + "grad_norm": 0.01519355084747076, + "learning_rate": 0.0006, + "loss": 4.114232063293457, + "step": 6694 + }, + { + "epoch": 92.99213973799127, + "grad_norm": 0.014596261084079742, + "learning_rate": 0.0006, + "loss": 3.992065906524658, + "step": 6695 + }, + { + "epoch": 93.0, + "grad_norm": 0.016378438100218773, + "learning_rate": 0.0006, + "loss": 4.093693733215332, + "step": 6696 + }, + { + "epoch": 93.0, + "eval_loss": 4.541905879974365, + "eval_runtime": 57.0177, + "eval_samples_per_second": 42.829, + "eval_steps_per_second": 1.35, + "step": 6696 + }, + { + "epoch": 93.01397379912663, + "grad_norm": 0.01593143492937088, + "learning_rate": 0.0006, + "loss": 4.052624225616455, + "step": 6697 + }, + { + "epoch": 93.02794759825328, + "grad_norm": 0.016086837276816368, + "learning_rate": 0.0006, + "loss": 4.130306243896484, + "step": 6698 + }, + { + "epoch": 93.04192139737991, + "grad_norm": 0.016004323959350586, + "learning_rate": 0.0006, + "loss": 3.965893268585205, + "step": 6699 + }, + { + "epoch": 93.05589519650655, + "grad_norm": 0.017667284235358238, + "learning_rate": 0.0006, + "loss": 4.095634460449219, + "step": 6700 + }, + { + "epoch": 93.06986899563319, + "grad_norm": 0.01949942484498024, + "learning_rate": 0.0006, + "loss": 3.966041326522827, + "step": 6701 + }, + { + "epoch": 93.08384279475983, + "grad_norm": 0.021898971870541573, + "learning_rate": 0.0006, + "loss": 4.010378837585449, + "step": 6702 + }, + { + "epoch": 93.09781659388646, + "grad_norm": 0.019985264167189598, + "learning_rate": 0.0006, + "loss": 4.061577320098877, + "step": 6703 + }, + { + "epoch": 93.1117903930131, + "grad_norm": 0.016470855101943016, + "learning_rate": 0.0006, + "loss": 4.0601677894592285, + "step": 6704 + }, + { + "epoch": 93.12576419213974, + "grad_norm": 0.015600379556417465, + "learning_rate": 0.0006, + "loss": 4.0860700607299805, + "step": 6705 + }, + { + "epoch": 93.13973799126637, + "grad_norm": 0.017514333128929138, + "learning_rate": 0.0006, + "loss": 4.240778923034668, + "step": 6706 + }, + { + "epoch": 93.15371179039302, + "grad_norm": 0.01780967228114605, + "learning_rate": 0.0006, + "loss": 4.076499938964844, + "step": 6707 + }, + { + "epoch": 93.16768558951965, + "grad_norm": 0.016370438039302826, + "learning_rate": 0.0006, + "loss": 4.0677618980407715, + "step": 6708 + }, + { + "epoch": 93.18165938864628, + "grad_norm": 0.015205491334199905, + "learning_rate": 0.0006, + "loss": 4.004945755004883, + "step": 6709 + }, + { + "epoch": 93.19563318777293, + "grad_norm": 0.01537297572940588, + "learning_rate": 0.0006, + "loss": 4.121147155761719, + "step": 6710 + }, + { + "epoch": 93.20960698689956, + "grad_norm": 0.017343711107969284, + "learning_rate": 0.0006, + "loss": 4.103215217590332, + "step": 6711 + }, + { + "epoch": 93.2235807860262, + "grad_norm": 0.01618131436407566, + "learning_rate": 0.0006, + "loss": 4.083741664886475, + "step": 6712 + }, + { + "epoch": 93.23755458515284, + "grad_norm": 0.013340235687792301, + "learning_rate": 0.0006, + "loss": 4.055032730102539, + "step": 6713 + }, + { + "epoch": 93.25152838427948, + "grad_norm": 0.014084120281040668, + "learning_rate": 0.0006, + "loss": 4.005610942840576, + "step": 6714 + }, + { + "epoch": 93.26550218340611, + "grad_norm": 0.016104530543088913, + "learning_rate": 0.0006, + "loss": 4.091764450073242, + "step": 6715 + }, + { + "epoch": 93.27947598253274, + "grad_norm": 0.015232563950121403, + "learning_rate": 0.0006, + "loss": 4.105735778808594, + "step": 6716 + }, + { + "epoch": 93.29344978165939, + "grad_norm": 0.014003508724272251, + "learning_rate": 0.0006, + "loss": 4.05783224105835, + "step": 6717 + }, + { + "epoch": 93.30742358078602, + "grad_norm": 0.01614833064377308, + "learning_rate": 0.0006, + "loss": 4.124567031860352, + "step": 6718 + }, + { + "epoch": 93.32139737991267, + "grad_norm": 0.015010425820946693, + "learning_rate": 0.0006, + "loss": 4.07656192779541, + "step": 6719 + }, + { + "epoch": 93.3353711790393, + "grad_norm": 0.01444012951105833, + "learning_rate": 0.0006, + "loss": 4.122415542602539, + "step": 6720 + }, + { + "epoch": 93.34934497816595, + "grad_norm": 0.015715785324573517, + "learning_rate": 0.0006, + "loss": 4.070283889770508, + "step": 6721 + }, + { + "epoch": 93.36331877729258, + "grad_norm": 0.014976450242102146, + "learning_rate": 0.0006, + "loss": 4.087043762207031, + "step": 6722 + }, + { + "epoch": 93.37729257641921, + "grad_norm": 0.014868560247123241, + "learning_rate": 0.0006, + "loss": 3.9671802520751953, + "step": 6723 + }, + { + "epoch": 93.39126637554585, + "grad_norm": 0.013518013060092926, + "learning_rate": 0.0006, + "loss": 4.136781692504883, + "step": 6724 + }, + { + "epoch": 93.40524017467249, + "grad_norm": 0.013498986139893532, + "learning_rate": 0.0006, + "loss": 3.982309103012085, + "step": 6725 + }, + { + "epoch": 93.41921397379913, + "grad_norm": 0.013707910664379597, + "learning_rate": 0.0006, + "loss": 4.181885719299316, + "step": 6726 + }, + { + "epoch": 93.43318777292576, + "grad_norm": 0.013913207687437534, + "learning_rate": 0.0006, + "loss": 3.980070114135742, + "step": 6727 + }, + { + "epoch": 93.44716157205241, + "grad_norm": 0.013169893994927406, + "learning_rate": 0.0006, + "loss": 4.023380279541016, + "step": 6728 + }, + { + "epoch": 93.46113537117904, + "grad_norm": 0.014451324008405209, + "learning_rate": 0.0006, + "loss": 4.038372039794922, + "step": 6729 + }, + { + "epoch": 93.47510917030567, + "grad_norm": 0.014401541091501713, + "learning_rate": 0.0006, + "loss": 4.069009304046631, + "step": 6730 + }, + { + "epoch": 93.48908296943232, + "grad_norm": 0.013971361331641674, + "learning_rate": 0.0006, + "loss": 4.105923652648926, + "step": 6731 + }, + { + "epoch": 93.50305676855895, + "grad_norm": 0.013192536309361458, + "learning_rate": 0.0006, + "loss": 4.07828950881958, + "step": 6732 + }, + { + "epoch": 93.5170305676856, + "grad_norm": 0.013581741601228714, + "learning_rate": 0.0006, + "loss": 4.079203128814697, + "step": 6733 + }, + { + "epoch": 93.53100436681223, + "grad_norm": 0.01508194301277399, + "learning_rate": 0.0006, + "loss": 4.084749221801758, + "step": 6734 + }, + { + "epoch": 93.54497816593886, + "grad_norm": 0.0156723503023386, + "learning_rate": 0.0006, + "loss": 4.070216178894043, + "step": 6735 + }, + { + "epoch": 93.5589519650655, + "grad_norm": 0.014996538870036602, + "learning_rate": 0.0006, + "loss": 4.111318588256836, + "step": 6736 + }, + { + "epoch": 93.57292576419214, + "grad_norm": 0.012815269641578197, + "learning_rate": 0.0006, + "loss": 4.130414962768555, + "step": 6737 + }, + { + "epoch": 93.58689956331878, + "grad_norm": 0.012887579388916492, + "learning_rate": 0.0006, + "loss": 4.128446102142334, + "step": 6738 + }, + { + "epoch": 93.60087336244541, + "grad_norm": 0.013632671907544136, + "learning_rate": 0.0006, + "loss": 4.20977258682251, + "step": 6739 + }, + { + "epoch": 93.61484716157206, + "grad_norm": 0.014318895526230335, + "learning_rate": 0.0006, + "loss": 4.005739212036133, + "step": 6740 + }, + { + "epoch": 93.62882096069869, + "grad_norm": 0.014160127379000187, + "learning_rate": 0.0006, + "loss": 4.122544288635254, + "step": 6741 + }, + { + "epoch": 93.64279475982532, + "grad_norm": 0.01308423187583685, + "learning_rate": 0.0006, + "loss": 4.108774662017822, + "step": 6742 + }, + { + "epoch": 93.65676855895197, + "grad_norm": 0.014293737709522247, + "learning_rate": 0.0006, + "loss": 3.983156681060791, + "step": 6743 + }, + { + "epoch": 93.6707423580786, + "grad_norm": 0.015155135653913021, + "learning_rate": 0.0006, + "loss": 4.006117820739746, + "step": 6744 + }, + { + "epoch": 93.68471615720524, + "grad_norm": 0.014391839504241943, + "learning_rate": 0.0006, + "loss": 4.1658935546875, + "step": 6745 + }, + { + "epoch": 93.69868995633188, + "grad_norm": 0.014924845658242702, + "learning_rate": 0.0006, + "loss": 4.090868949890137, + "step": 6746 + }, + { + "epoch": 93.71266375545852, + "grad_norm": 0.014786627143621445, + "learning_rate": 0.0006, + "loss": 4.121716499328613, + "step": 6747 + }, + { + "epoch": 93.72663755458515, + "grad_norm": 0.016948824748396873, + "learning_rate": 0.0006, + "loss": 4.213179111480713, + "step": 6748 + }, + { + "epoch": 93.74061135371178, + "grad_norm": 0.01732620596885681, + "learning_rate": 0.0006, + "loss": 4.079215049743652, + "step": 6749 + }, + { + "epoch": 93.75458515283843, + "grad_norm": 0.016298124566674232, + "learning_rate": 0.0006, + "loss": 4.096126079559326, + "step": 6750 + }, + { + "epoch": 93.76855895196506, + "grad_norm": 0.015744732692837715, + "learning_rate": 0.0006, + "loss": 4.046300411224365, + "step": 6751 + }, + { + "epoch": 93.78253275109171, + "grad_norm": 0.016330812126398087, + "learning_rate": 0.0006, + "loss": 4.098011016845703, + "step": 6752 + }, + { + "epoch": 93.79650655021834, + "grad_norm": 0.01710600219666958, + "learning_rate": 0.0006, + "loss": 4.031402587890625, + "step": 6753 + }, + { + "epoch": 93.81048034934499, + "grad_norm": 0.01561805047094822, + "learning_rate": 0.0006, + "loss": 4.023388862609863, + "step": 6754 + }, + { + "epoch": 93.82445414847162, + "grad_norm": 0.013637243770062923, + "learning_rate": 0.0006, + "loss": 4.143238067626953, + "step": 6755 + }, + { + "epoch": 93.83842794759825, + "grad_norm": 0.013088075444102287, + "learning_rate": 0.0006, + "loss": 3.9083704948425293, + "step": 6756 + }, + { + "epoch": 93.8524017467249, + "grad_norm": 0.015893537551164627, + "learning_rate": 0.0006, + "loss": 4.102263450622559, + "step": 6757 + }, + { + "epoch": 93.86637554585153, + "grad_norm": 0.015213928185403347, + "learning_rate": 0.0006, + "loss": 4.111978530883789, + "step": 6758 + }, + { + "epoch": 93.88034934497817, + "grad_norm": 0.014543982222676277, + "learning_rate": 0.0006, + "loss": 4.05889892578125, + "step": 6759 + }, + { + "epoch": 93.8943231441048, + "grad_norm": 0.015104752965271473, + "learning_rate": 0.0006, + "loss": 4.136166572570801, + "step": 6760 + }, + { + "epoch": 93.90829694323143, + "grad_norm": 0.01505881268531084, + "learning_rate": 0.0006, + "loss": 4.061883449554443, + "step": 6761 + }, + { + "epoch": 93.92227074235808, + "grad_norm": 0.015248455107212067, + "learning_rate": 0.0006, + "loss": 4.109863758087158, + "step": 6762 + }, + { + "epoch": 93.93624454148471, + "grad_norm": 0.013651185669004917, + "learning_rate": 0.0006, + "loss": 4.061549663543701, + "step": 6763 + }, + { + "epoch": 93.95021834061136, + "grad_norm": 0.013806282542645931, + "learning_rate": 0.0006, + "loss": 4.205403804779053, + "step": 6764 + }, + { + "epoch": 93.96419213973799, + "grad_norm": 0.01538451761007309, + "learning_rate": 0.0006, + "loss": 4.14816951751709, + "step": 6765 + }, + { + "epoch": 93.97816593886463, + "grad_norm": 0.01602427288889885, + "learning_rate": 0.0006, + "loss": 4.000303268432617, + "step": 6766 + }, + { + "epoch": 93.99213973799127, + "grad_norm": 0.014558793976902962, + "learning_rate": 0.0006, + "loss": 4.096562385559082, + "step": 6767 + }, + { + "epoch": 94.0, + "grad_norm": 0.01613018475472927, + "learning_rate": 0.0006, + "loss": 4.262835502624512, + "step": 6768 + }, + { + "epoch": 94.0, + "eval_loss": 4.532126426696777, + "eval_runtime": 56.7592, + "eval_samples_per_second": 43.024, + "eval_steps_per_second": 1.357, + "step": 6768 + }, + { + "epoch": 94.01397379912663, + "grad_norm": 0.01614535041153431, + "learning_rate": 0.0006, + "loss": 4.073431491851807, + "step": 6769 + }, + { + "epoch": 94.02794759825328, + "grad_norm": 0.015986818820238113, + "learning_rate": 0.0006, + "loss": 4.070001602172852, + "step": 6770 + }, + { + "epoch": 94.04192139737991, + "grad_norm": 0.015678714960813522, + "learning_rate": 0.0006, + "loss": 4.06920051574707, + "step": 6771 + }, + { + "epoch": 94.05589519650655, + "grad_norm": 0.016593851149082184, + "learning_rate": 0.0006, + "loss": 4.095780849456787, + "step": 6772 + }, + { + "epoch": 94.06986899563319, + "grad_norm": 0.01780948042869568, + "learning_rate": 0.0006, + "loss": 4.092916488647461, + "step": 6773 + }, + { + "epoch": 94.08384279475983, + "grad_norm": 0.01755528897047043, + "learning_rate": 0.0006, + "loss": 3.9833455085754395, + "step": 6774 + }, + { + "epoch": 94.09781659388646, + "grad_norm": 0.01708311215043068, + "learning_rate": 0.0006, + "loss": 4.053336143493652, + "step": 6775 + }, + { + "epoch": 94.1117903930131, + "grad_norm": 0.017362922430038452, + "learning_rate": 0.0006, + "loss": 4.014343738555908, + "step": 6776 + }, + { + "epoch": 94.12576419213974, + "grad_norm": 0.01824064739048481, + "learning_rate": 0.0006, + "loss": 4.1215691566467285, + "step": 6777 + }, + { + "epoch": 94.13973799126637, + "grad_norm": 0.017743906006217003, + "learning_rate": 0.0006, + "loss": 4.091215133666992, + "step": 6778 + }, + { + "epoch": 94.15371179039302, + "grad_norm": 0.016585713252425194, + "learning_rate": 0.0006, + "loss": 4.013662338256836, + "step": 6779 + }, + { + "epoch": 94.16768558951965, + "grad_norm": 0.017397213727235794, + "learning_rate": 0.0006, + "loss": 4.070602893829346, + "step": 6780 + }, + { + "epoch": 94.18165938864628, + "grad_norm": 0.01733497530221939, + "learning_rate": 0.0006, + "loss": 4.1304426193237305, + "step": 6781 + }, + { + "epoch": 94.19563318777293, + "grad_norm": 0.015215300023555756, + "learning_rate": 0.0006, + "loss": 3.965014696121216, + "step": 6782 + }, + { + "epoch": 94.20960698689956, + "grad_norm": 0.015908699482679367, + "learning_rate": 0.0006, + "loss": 3.924964427947998, + "step": 6783 + }, + { + "epoch": 94.2235807860262, + "grad_norm": 0.014044968411326408, + "learning_rate": 0.0006, + "loss": 4.08673095703125, + "step": 6784 + }, + { + "epoch": 94.23755458515284, + "grad_norm": 0.01375142764300108, + "learning_rate": 0.0006, + "loss": 4.011466979980469, + "step": 6785 + }, + { + "epoch": 94.25152838427948, + "grad_norm": 0.015539257787168026, + "learning_rate": 0.0006, + "loss": 4.09976863861084, + "step": 6786 + }, + { + "epoch": 94.26550218340611, + "grad_norm": 0.014303367584943771, + "learning_rate": 0.0006, + "loss": 4.108369827270508, + "step": 6787 + }, + { + "epoch": 94.27947598253274, + "grad_norm": 0.01281928364187479, + "learning_rate": 0.0006, + "loss": 4.073737144470215, + "step": 6788 + }, + { + "epoch": 94.29344978165939, + "grad_norm": 0.013859959319233894, + "learning_rate": 0.0006, + "loss": 4.054823875427246, + "step": 6789 + }, + { + "epoch": 94.30742358078602, + "grad_norm": 0.013326860964298248, + "learning_rate": 0.0006, + "loss": 4.133251190185547, + "step": 6790 + }, + { + "epoch": 94.32139737991267, + "grad_norm": 0.013923076912760735, + "learning_rate": 0.0006, + "loss": 4.190706729888916, + "step": 6791 + }, + { + "epoch": 94.3353711790393, + "grad_norm": 0.01420847699046135, + "learning_rate": 0.0006, + "loss": 3.9336605072021484, + "step": 6792 + }, + { + "epoch": 94.34934497816595, + "grad_norm": 0.014607555232942104, + "learning_rate": 0.0006, + "loss": 4.128399848937988, + "step": 6793 + }, + { + "epoch": 94.36331877729258, + "grad_norm": 0.012862961739301682, + "learning_rate": 0.0006, + "loss": 4.0132527351379395, + "step": 6794 + }, + { + "epoch": 94.37729257641921, + "grad_norm": 0.012530217878520489, + "learning_rate": 0.0006, + "loss": 4.0384368896484375, + "step": 6795 + }, + { + "epoch": 94.39126637554585, + "grad_norm": 0.01286802627146244, + "learning_rate": 0.0006, + "loss": 3.95570707321167, + "step": 6796 + }, + { + "epoch": 94.40524017467249, + "grad_norm": 0.013450917787849903, + "learning_rate": 0.0006, + "loss": 4.016484260559082, + "step": 6797 + }, + { + "epoch": 94.41921397379913, + "grad_norm": 0.01393423043191433, + "learning_rate": 0.0006, + "loss": 4.124959468841553, + "step": 6798 + }, + { + "epoch": 94.43318777292576, + "grad_norm": 0.013077184557914734, + "learning_rate": 0.0006, + "loss": 4.029163837432861, + "step": 6799 + }, + { + "epoch": 94.44716157205241, + "grad_norm": 0.013577334582805634, + "learning_rate": 0.0006, + "loss": 4.22929573059082, + "step": 6800 + }, + { + "epoch": 94.46113537117904, + "grad_norm": 0.013930312357842922, + "learning_rate": 0.0006, + "loss": 4.024711608886719, + "step": 6801 + }, + { + "epoch": 94.47510917030567, + "grad_norm": 0.013746250420808792, + "learning_rate": 0.0006, + "loss": 3.9766428470611572, + "step": 6802 + }, + { + "epoch": 94.48908296943232, + "grad_norm": 0.013682983815670013, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6803 + }, + { + "epoch": 94.50305676855895, + "grad_norm": 0.01473842654377222, + "learning_rate": 0.0006, + "loss": 4.197623252868652, + "step": 6804 + }, + { + "epoch": 94.5170305676856, + "grad_norm": 0.014133758842945099, + "learning_rate": 0.0006, + "loss": 4.055549621582031, + "step": 6805 + }, + { + "epoch": 94.53100436681223, + "grad_norm": 0.0139645179733634, + "learning_rate": 0.0006, + "loss": 4.130135536193848, + "step": 6806 + }, + { + "epoch": 94.54497816593886, + "grad_norm": 0.013973365537822247, + "learning_rate": 0.0006, + "loss": 4.006230354309082, + "step": 6807 + }, + { + "epoch": 94.5589519650655, + "grad_norm": 0.014227126725018024, + "learning_rate": 0.0006, + "loss": 4.130459785461426, + "step": 6808 + }, + { + "epoch": 94.57292576419214, + "grad_norm": 0.013129506260156631, + "learning_rate": 0.0006, + "loss": 4.014974594116211, + "step": 6809 + }, + { + "epoch": 94.58689956331878, + "grad_norm": 0.013015891425311565, + "learning_rate": 0.0006, + "loss": 4.161370277404785, + "step": 6810 + }, + { + "epoch": 94.60087336244541, + "grad_norm": 0.013957147486507893, + "learning_rate": 0.0006, + "loss": 4.09519100189209, + "step": 6811 + }, + { + "epoch": 94.61484716157206, + "grad_norm": 0.015494572930037975, + "learning_rate": 0.0006, + "loss": 4.029949188232422, + "step": 6812 + }, + { + "epoch": 94.62882096069869, + "grad_norm": 0.016732869669795036, + "learning_rate": 0.0006, + "loss": 4.135669708251953, + "step": 6813 + }, + { + "epoch": 94.64279475982532, + "grad_norm": 0.01769322343170643, + "learning_rate": 0.0006, + "loss": 4.109872817993164, + "step": 6814 + }, + { + "epoch": 94.65676855895197, + "grad_norm": 0.015932537615299225, + "learning_rate": 0.0006, + "loss": 4.014595031738281, + "step": 6815 + }, + { + "epoch": 94.6707423580786, + "grad_norm": 0.015152990818023682, + "learning_rate": 0.0006, + "loss": 4.1983442306518555, + "step": 6816 + }, + { + "epoch": 94.68471615720524, + "grad_norm": 0.01590685173869133, + "learning_rate": 0.0006, + "loss": 4.119250774383545, + "step": 6817 + }, + { + "epoch": 94.69868995633188, + "grad_norm": 0.016018465161323547, + "learning_rate": 0.0006, + "loss": 4.048069953918457, + "step": 6818 + }, + { + "epoch": 94.71266375545852, + "grad_norm": 0.016621652990579605, + "learning_rate": 0.0006, + "loss": 4.043731689453125, + "step": 6819 + }, + { + "epoch": 94.72663755458515, + "grad_norm": 0.014535024762153625, + "learning_rate": 0.0006, + "loss": 4.1886820793151855, + "step": 6820 + }, + { + "epoch": 94.74061135371178, + "grad_norm": 0.014192107133567333, + "learning_rate": 0.0006, + "loss": 4.103938102722168, + "step": 6821 + }, + { + "epoch": 94.75458515283843, + "grad_norm": 0.0166022852063179, + "learning_rate": 0.0006, + "loss": 4.017818450927734, + "step": 6822 + }, + { + "epoch": 94.76855895196506, + "grad_norm": 0.016247155144810677, + "learning_rate": 0.0006, + "loss": 4.0038981437683105, + "step": 6823 + }, + { + "epoch": 94.78253275109171, + "grad_norm": 0.01593470573425293, + "learning_rate": 0.0006, + "loss": 4.043969631195068, + "step": 6824 + }, + { + "epoch": 94.79650655021834, + "grad_norm": 0.014334345236420631, + "learning_rate": 0.0006, + "loss": 4.097829341888428, + "step": 6825 + }, + { + "epoch": 94.81048034934499, + "grad_norm": 0.015581986866891384, + "learning_rate": 0.0006, + "loss": 4.148984432220459, + "step": 6826 + }, + { + "epoch": 94.82445414847162, + "grad_norm": 0.017519738525152206, + "learning_rate": 0.0006, + "loss": 4.245623588562012, + "step": 6827 + }, + { + "epoch": 94.83842794759825, + "grad_norm": 0.015155277214944363, + "learning_rate": 0.0006, + "loss": 4.031019687652588, + "step": 6828 + }, + { + "epoch": 94.8524017467249, + "grad_norm": 0.014969222247600555, + "learning_rate": 0.0006, + "loss": 4.036084175109863, + "step": 6829 + }, + { + "epoch": 94.86637554585153, + "grad_norm": 0.015987908467650414, + "learning_rate": 0.0006, + "loss": 4.08396053314209, + "step": 6830 + }, + { + "epoch": 94.88034934497817, + "grad_norm": 0.016324151307344437, + "learning_rate": 0.0006, + "loss": 4.036787033081055, + "step": 6831 + }, + { + "epoch": 94.8943231441048, + "grad_norm": 0.015957487747073174, + "learning_rate": 0.0006, + "loss": 4.147607803344727, + "step": 6832 + }, + { + "epoch": 94.90829694323143, + "grad_norm": 0.015286967158317566, + "learning_rate": 0.0006, + "loss": 4.184615135192871, + "step": 6833 + }, + { + "epoch": 94.92227074235808, + "grad_norm": 0.01338377594947815, + "learning_rate": 0.0006, + "loss": 4.014981269836426, + "step": 6834 + }, + { + "epoch": 94.93624454148471, + "grad_norm": 0.013997633941471577, + "learning_rate": 0.0006, + "loss": 4.021933078765869, + "step": 6835 + }, + { + "epoch": 94.95021834061136, + "grad_norm": 0.014572402462363243, + "learning_rate": 0.0006, + "loss": 4.0326948165893555, + "step": 6836 + }, + { + "epoch": 94.96419213973799, + "grad_norm": 0.014379586093127728, + "learning_rate": 0.0006, + "loss": 4.060697555541992, + "step": 6837 + }, + { + "epoch": 94.97816593886463, + "grad_norm": 0.013513042591512203, + "learning_rate": 0.0006, + "loss": 4.05123233795166, + "step": 6838 + }, + { + "epoch": 94.99213973799127, + "grad_norm": 0.012904108501970768, + "learning_rate": 0.0006, + "loss": 4.020852565765381, + "step": 6839 + }, + { + "epoch": 95.0, + "grad_norm": 0.014364209026098251, + "learning_rate": 0.0006, + "loss": 4.080837249755859, + "step": 6840 + }, + { + "epoch": 95.0, + "eval_loss": 4.608776569366455, + "eval_runtime": 56.6516, + "eval_samples_per_second": 43.106, + "eval_steps_per_second": 1.359, + "step": 6840 + }, + { + "epoch": 95.01397379912663, + "grad_norm": 0.01357510220259428, + "learning_rate": 0.0006, + "loss": 4.105312347412109, + "step": 6841 + }, + { + "epoch": 95.02794759825328, + "grad_norm": 0.014248257502913475, + "learning_rate": 0.0006, + "loss": 3.971667766571045, + "step": 6842 + }, + { + "epoch": 95.04192139737991, + "grad_norm": 0.014773483388125896, + "learning_rate": 0.0006, + "loss": 4.03553581237793, + "step": 6843 + }, + { + "epoch": 95.05589519650655, + "grad_norm": 0.014766111969947815, + "learning_rate": 0.0006, + "loss": 4.074707508087158, + "step": 6844 + }, + { + "epoch": 95.06986899563319, + "grad_norm": 0.014778018929064274, + "learning_rate": 0.0006, + "loss": 4.1083855628967285, + "step": 6845 + }, + { + "epoch": 95.08384279475983, + "grad_norm": 0.014466686174273491, + "learning_rate": 0.0006, + "loss": 4.054587364196777, + "step": 6846 + }, + { + "epoch": 95.09781659388646, + "grad_norm": 0.015195904299616814, + "learning_rate": 0.0006, + "loss": 4.142580032348633, + "step": 6847 + }, + { + "epoch": 95.1117903930131, + "grad_norm": 0.01530557032674551, + "learning_rate": 0.0006, + "loss": 4.062134265899658, + "step": 6848 + }, + { + "epoch": 95.12576419213974, + "grad_norm": 0.015203768387436867, + "learning_rate": 0.0006, + "loss": 4.082962989807129, + "step": 6849 + }, + { + "epoch": 95.13973799126637, + "grad_norm": 0.014906628057360649, + "learning_rate": 0.0006, + "loss": 4.033127784729004, + "step": 6850 + }, + { + "epoch": 95.15371179039302, + "grad_norm": 0.01553534809499979, + "learning_rate": 0.0006, + "loss": 4.051743030548096, + "step": 6851 + }, + { + "epoch": 95.16768558951965, + "grad_norm": 0.01607976108789444, + "learning_rate": 0.0006, + "loss": 4.038073539733887, + "step": 6852 + }, + { + "epoch": 95.18165938864628, + "grad_norm": 0.015776360407471657, + "learning_rate": 0.0006, + "loss": 4.013091087341309, + "step": 6853 + }, + { + "epoch": 95.19563318777293, + "grad_norm": 0.01623626798391342, + "learning_rate": 0.0006, + "loss": 4.12293815612793, + "step": 6854 + }, + { + "epoch": 95.20960698689956, + "grad_norm": 0.016887828707695007, + "learning_rate": 0.0006, + "loss": 4.120452404022217, + "step": 6855 + }, + { + "epoch": 95.2235807860262, + "grad_norm": 0.01561128068715334, + "learning_rate": 0.0006, + "loss": 4.071322441101074, + "step": 6856 + }, + { + "epoch": 95.23755458515284, + "grad_norm": 0.015679724514484406, + "learning_rate": 0.0006, + "loss": 4.065072059631348, + "step": 6857 + }, + { + "epoch": 95.25152838427948, + "grad_norm": 0.016506962478160858, + "learning_rate": 0.0006, + "loss": 4.0667924880981445, + "step": 6858 + }, + { + "epoch": 95.26550218340611, + "grad_norm": 0.015340839512646198, + "learning_rate": 0.0006, + "loss": 4.14047908782959, + "step": 6859 + }, + { + "epoch": 95.27947598253274, + "grad_norm": 0.013935413211584091, + "learning_rate": 0.0006, + "loss": 3.9688191413879395, + "step": 6860 + }, + { + "epoch": 95.29344978165939, + "grad_norm": 0.015223405323922634, + "learning_rate": 0.0006, + "loss": 4.101613521575928, + "step": 6861 + }, + { + "epoch": 95.30742358078602, + "grad_norm": 0.01562546379864216, + "learning_rate": 0.0006, + "loss": 4.128626823425293, + "step": 6862 + }, + { + "epoch": 95.32139737991267, + "grad_norm": 0.016159312799572945, + "learning_rate": 0.0006, + "loss": 4.080667018890381, + "step": 6863 + }, + { + "epoch": 95.3353711790393, + "grad_norm": 0.01724511943757534, + "learning_rate": 0.0006, + "loss": 4.08463191986084, + "step": 6864 + }, + { + "epoch": 95.34934497816595, + "grad_norm": 0.01592276245355606, + "learning_rate": 0.0006, + "loss": 4.127612113952637, + "step": 6865 + }, + { + "epoch": 95.36331877729258, + "grad_norm": 0.01744082197546959, + "learning_rate": 0.0006, + "loss": 4.034480094909668, + "step": 6866 + }, + { + "epoch": 95.37729257641921, + "grad_norm": 0.01774069108068943, + "learning_rate": 0.0006, + "loss": 4.091196537017822, + "step": 6867 + }, + { + "epoch": 95.39126637554585, + "grad_norm": 0.01653675176203251, + "learning_rate": 0.0006, + "loss": 4.032937526702881, + "step": 6868 + }, + { + "epoch": 95.40524017467249, + "grad_norm": 0.014933439902961254, + "learning_rate": 0.0006, + "loss": 3.9386508464813232, + "step": 6869 + }, + { + "epoch": 95.41921397379913, + "grad_norm": 0.014743257313966751, + "learning_rate": 0.0006, + "loss": 4.126462459564209, + "step": 6870 + }, + { + "epoch": 95.43318777292576, + "grad_norm": 0.014614643529057503, + "learning_rate": 0.0006, + "loss": 4.093838214874268, + "step": 6871 + }, + { + "epoch": 95.44716157205241, + "grad_norm": 0.01665084809064865, + "learning_rate": 0.0006, + "loss": 4.1471967697143555, + "step": 6872 + }, + { + "epoch": 95.46113537117904, + "grad_norm": 0.018056001514196396, + "learning_rate": 0.0006, + "loss": 4.1388349533081055, + "step": 6873 + }, + { + "epoch": 95.47510917030567, + "grad_norm": 0.02129282057285309, + "learning_rate": 0.0006, + "loss": 4.130726337432861, + "step": 6874 + }, + { + "epoch": 95.48908296943232, + "grad_norm": 0.022237276658415794, + "learning_rate": 0.0006, + "loss": 4.10750675201416, + "step": 6875 + }, + { + "epoch": 95.50305676855895, + "grad_norm": 0.021106960251927376, + "learning_rate": 0.0006, + "loss": 4.136451721191406, + "step": 6876 + }, + { + "epoch": 95.5170305676856, + "grad_norm": 0.019107872620224953, + "learning_rate": 0.0006, + "loss": 4.139644145965576, + "step": 6877 + }, + { + "epoch": 95.53100436681223, + "grad_norm": 0.018290145322680473, + "learning_rate": 0.0006, + "loss": 4.161884784698486, + "step": 6878 + }, + { + "epoch": 95.54497816593886, + "grad_norm": 0.018161306157708168, + "learning_rate": 0.0006, + "loss": 4.1252336502075195, + "step": 6879 + }, + { + "epoch": 95.5589519650655, + "grad_norm": 0.018432756885886192, + "learning_rate": 0.0006, + "loss": 4.0635986328125, + "step": 6880 + }, + { + "epoch": 95.57292576419214, + "grad_norm": 0.016592558473348618, + "learning_rate": 0.0006, + "loss": 4.037112236022949, + "step": 6881 + }, + { + "epoch": 95.58689956331878, + "grad_norm": 0.015283452346920967, + "learning_rate": 0.0006, + "loss": 4.1964826583862305, + "step": 6882 + }, + { + "epoch": 95.60087336244541, + "grad_norm": 0.015539180487394333, + "learning_rate": 0.0006, + "loss": 4.110917568206787, + "step": 6883 + }, + { + "epoch": 95.61484716157206, + "grad_norm": 0.016508113592863083, + "learning_rate": 0.0006, + "loss": 3.9705517292022705, + "step": 6884 + }, + { + "epoch": 95.62882096069869, + "grad_norm": 0.016635406762361526, + "learning_rate": 0.0006, + "loss": 4.025636196136475, + "step": 6885 + }, + { + "epoch": 95.64279475982532, + "grad_norm": 0.01688900962471962, + "learning_rate": 0.0006, + "loss": 4.01814079284668, + "step": 6886 + }, + { + "epoch": 95.65676855895197, + "grad_norm": 0.01614062674343586, + "learning_rate": 0.0006, + "loss": 4.098041534423828, + "step": 6887 + }, + { + "epoch": 95.6707423580786, + "grad_norm": 0.015679053962230682, + "learning_rate": 0.0006, + "loss": 4.077049255371094, + "step": 6888 + }, + { + "epoch": 95.68471615720524, + "grad_norm": 0.016617605462670326, + "learning_rate": 0.0006, + "loss": 4.10363245010376, + "step": 6889 + }, + { + "epoch": 95.69868995633188, + "grad_norm": 0.015966709703207016, + "learning_rate": 0.0006, + "loss": 4.100924491882324, + "step": 6890 + }, + { + "epoch": 95.71266375545852, + "grad_norm": 0.015768567100167274, + "learning_rate": 0.0006, + "loss": 4.063976287841797, + "step": 6891 + }, + { + "epoch": 95.72663755458515, + "grad_norm": 0.015800341963768005, + "learning_rate": 0.0006, + "loss": 4.008409023284912, + "step": 6892 + }, + { + "epoch": 95.74061135371178, + "grad_norm": 0.014288571663200855, + "learning_rate": 0.0006, + "loss": 4.139591217041016, + "step": 6893 + }, + { + "epoch": 95.75458515283843, + "grad_norm": 0.013743448071181774, + "learning_rate": 0.0006, + "loss": 3.967352867126465, + "step": 6894 + }, + { + "epoch": 95.76855895196506, + "grad_norm": 0.01387989055365324, + "learning_rate": 0.0006, + "loss": 4.127554893493652, + "step": 6895 + }, + { + "epoch": 95.78253275109171, + "grad_norm": 0.013910802081227303, + "learning_rate": 0.0006, + "loss": 4.130941390991211, + "step": 6896 + }, + { + "epoch": 95.79650655021834, + "grad_norm": 0.013936805538833141, + "learning_rate": 0.0006, + "loss": 4.138744831085205, + "step": 6897 + }, + { + "epoch": 95.81048034934499, + "grad_norm": 0.015135723166167736, + "learning_rate": 0.0006, + "loss": 4.131860256195068, + "step": 6898 + }, + { + "epoch": 95.82445414847162, + "grad_norm": 0.014921358786523342, + "learning_rate": 0.0006, + "loss": 4.1949462890625, + "step": 6899 + }, + { + "epoch": 95.83842794759825, + "grad_norm": 0.01401488296687603, + "learning_rate": 0.0006, + "loss": 4.181267738342285, + "step": 6900 + }, + { + "epoch": 95.8524017467249, + "grad_norm": 0.013691626489162445, + "learning_rate": 0.0006, + "loss": 4.113667964935303, + "step": 6901 + }, + { + "epoch": 95.86637554585153, + "grad_norm": 0.013890047557651997, + "learning_rate": 0.0006, + "loss": 4.163067817687988, + "step": 6902 + }, + { + "epoch": 95.88034934497817, + "grad_norm": 0.01385503076016903, + "learning_rate": 0.0006, + "loss": 4.076153755187988, + "step": 6903 + }, + { + "epoch": 95.8943231441048, + "grad_norm": 0.01383926346898079, + "learning_rate": 0.0006, + "loss": 4.102423667907715, + "step": 6904 + }, + { + "epoch": 95.90829694323143, + "grad_norm": 0.015961717814207077, + "learning_rate": 0.0006, + "loss": 4.1543779373168945, + "step": 6905 + }, + { + "epoch": 95.92227074235808, + "grad_norm": 0.016718363389372826, + "learning_rate": 0.0006, + "loss": 3.991607189178467, + "step": 6906 + }, + { + "epoch": 95.93624454148471, + "grad_norm": 0.01585792563855648, + "learning_rate": 0.0006, + "loss": 4.048587799072266, + "step": 6907 + }, + { + "epoch": 95.95021834061136, + "grad_norm": 0.015311713330447674, + "learning_rate": 0.0006, + "loss": 4.1070451736450195, + "step": 6908 + }, + { + "epoch": 95.96419213973799, + "grad_norm": 0.014521263539791107, + "learning_rate": 0.0006, + "loss": 4.087587356567383, + "step": 6909 + }, + { + "epoch": 95.97816593886463, + "grad_norm": 0.01433157455176115, + "learning_rate": 0.0006, + "loss": 4.139904499053955, + "step": 6910 + }, + { + "epoch": 95.99213973799127, + "grad_norm": 0.014475514180958271, + "learning_rate": 0.0006, + "loss": 4.182525634765625, + "step": 6911 + }, + { + "epoch": 96.0, + "grad_norm": 0.015245620161294937, + "learning_rate": 0.0006, + "loss": 4.1199870109558105, + "step": 6912 + }, + { + "epoch": 96.0, + "eval_loss": 4.58721399307251, + "eval_runtime": 56.395, + "eval_samples_per_second": 43.302, + "eval_steps_per_second": 1.365, + "step": 6912 + }, + { + "epoch": 96.01397379912663, + "grad_norm": 0.014325885102152824, + "learning_rate": 0.0006, + "loss": 4.025214195251465, + "step": 6913 + }, + { + "epoch": 96.02794759825328, + "grad_norm": 0.01678297482430935, + "learning_rate": 0.0006, + "loss": 3.998842716217041, + "step": 6914 + }, + { + "epoch": 96.04192139737991, + "grad_norm": 0.017044762149453163, + "learning_rate": 0.0006, + "loss": 4.122270584106445, + "step": 6915 + }, + { + "epoch": 96.05589519650655, + "grad_norm": 0.017583075910806656, + "learning_rate": 0.0006, + "loss": 3.9782891273498535, + "step": 6916 + }, + { + "epoch": 96.06986899563319, + "grad_norm": 0.0179787315428257, + "learning_rate": 0.0006, + "loss": 4.030402660369873, + "step": 6917 + }, + { + "epoch": 96.08384279475983, + "grad_norm": 0.0159553661942482, + "learning_rate": 0.0006, + "loss": 3.949040651321411, + "step": 6918 + }, + { + "epoch": 96.09781659388646, + "grad_norm": 0.01554026547819376, + "learning_rate": 0.0006, + "loss": 4.0429534912109375, + "step": 6919 + }, + { + "epoch": 96.1117903930131, + "grad_norm": 0.016099359840154648, + "learning_rate": 0.0006, + "loss": 4.037537097930908, + "step": 6920 + }, + { + "epoch": 96.12576419213974, + "grad_norm": 0.01676846854388714, + "learning_rate": 0.0006, + "loss": 4.022023677825928, + "step": 6921 + }, + { + "epoch": 96.13973799126637, + "grad_norm": 0.019358212128281593, + "learning_rate": 0.0006, + "loss": 4.048679351806641, + "step": 6922 + }, + { + "epoch": 96.15371179039302, + "grad_norm": 0.020390303805470467, + "learning_rate": 0.0006, + "loss": 4.144132614135742, + "step": 6923 + }, + { + "epoch": 96.16768558951965, + "grad_norm": 0.01884407177567482, + "learning_rate": 0.0006, + "loss": 4.128715515136719, + "step": 6924 + }, + { + "epoch": 96.18165938864628, + "grad_norm": 0.015098451636731625, + "learning_rate": 0.0006, + "loss": 4.128741264343262, + "step": 6925 + }, + { + "epoch": 96.19563318777293, + "grad_norm": 0.0148141635581851, + "learning_rate": 0.0006, + "loss": 4.054587364196777, + "step": 6926 + }, + { + "epoch": 96.20960698689956, + "grad_norm": 0.015594304539263248, + "learning_rate": 0.0006, + "loss": 4.073890209197998, + "step": 6927 + }, + { + "epoch": 96.2235807860262, + "grad_norm": 0.0139383003115654, + "learning_rate": 0.0006, + "loss": 4.115614891052246, + "step": 6928 + }, + { + "epoch": 96.23755458515284, + "grad_norm": 0.0150460135191679, + "learning_rate": 0.0006, + "loss": 4.040738105773926, + "step": 6929 + }, + { + "epoch": 96.25152838427948, + "grad_norm": 0.013879122212529182, + "learning_rate": 0.0006, + "loss": 3.9292469024658203, + "step": 6930 + }, + { + "epoch": 96.26550218340611, + "grad_norm": 0.015546537935733795, + "learning_rate": 0.0006, + "loss": 4.111361503601074, + "step": 6931 + }, + { + "epoch": 96.27947598253274, + "grad_norm": 0.016172925010323524, + "learning_rate": 0.0006, + "loss": 4.041029930114746, + "step": 6932 + }, + { + "epoch": 96.29344978165939, + "grad_norm": 0.015113193541765213, + "learning_rate": 0.0006, + "loss": 4.157950401306152, + "step": 6933 + }, + { + "epoch": 96.30742358078602, + "grad_norm": 0.014663247391581535, + "learning_rate": 0.0006, + "loss": 4.020952224731445, + "step": 6934 + }, + { + "epoch": 96.32139737991267, + "grad_norm": 0.015092858113348484, + "learning_rate": 0.0006, + "loss": 4.065241813659668, + "step": 6935 + }, + { + "epoch": 96.3353711790393, + "grad_norm": 0.013199460692703724, + "learning_rate": 0.0006, + "loss": 4.062674522399902, + "step": 6936 + }, + { + "epoch": 96.34934497816595, + "grad_norm": 0.015065675601363182, + "learning_rate": 0.0006, + "loss": 4.038748741149902, + "step": 6937 + }, + { + "epoch": 96.36331877729258, + "grad_norm": 0.014270043931901455, + "learning_rate": 0.0006, + "loss": 4.073092460632324, + "step": 6938 + }, + { + "epoch": 96.37729257641921, + "grad_norm": 0.01423644833266735, + "learning_rate": 0.0006, + "loss": 4.238361835479736, + "step": 6939 + }, + { + "epoch": 96.39126637554585, + "grad_norm": 0.01380224246531725, + "learning_rate": 0.0006, + "loss": 4.137866973876953, + "step": 6940 + }, + { + "epoch": 96.40524017467249, + "grad_norm": 0.012464135885238647, + "learning_rate": 0.0006, + "loss": 3.9527339935302734, + "step": 6941 + }, + { + "epoch": 96.41921397379913, + "grad_norm": 0.013094968162477016, + "learning_rate": 0.0006, + "loss": 4.106327056884766, + "step": 6942 + }, + { + "epoch": 96.43318777292576, + "grad_norm": 0.013918030075728893, + "learning_rate": 0.0006, + "loss": 4.072292327880859, + "step": 6943 + }, + { + "epoch": 96.44716157205241, + "grad_norm": 0.013835827820003033, + "learning_rate": 0.0006, + "loss": 4.094313621520996, + "step": 6944 + }, + { + "epoch": 96.46113537117904, + "grad_norm": 0.013890027068555355, + "learning_rate": 0.0006, + "loss": 3.947615623474121, + "step": 6945 + }, + { + "epoch": 96.47510917030567, + "grad_norm": 0.012371144257485867, + "learning_rate": 0.0006, + "loss": 4.038928985595703, + "step": 6946 + }, + { + "epoch": 96.48908296943232, + "grad_norm": 0.012236416339874268, + "learning_rate": 0.0006, + "loss": 3.9835925102233887, + "step": 6947 + }, + { + "epoch": 96.50305676855895, + "grad_norm": 0.012866619043052197, + "learning_rate": 0.0006, + "loss": 4.076041221618652, + "step": 6948 + }, + { + "epoch": 96.5170305676856, + "grad_norm": 0.013425248675048351, + "learning_rate": 0.0006, + "loss": 4.086307525634766, + "step": 6949 + }, + { + "epoch": 96.53100436681223, + "grad_norm": 0.014234314672648907, + "learning_rate": 0.0006, + "loss": 4.087897300720215, + "step": 6950 + }, + { + "epoch": 96.54497816593886, + "grad_norm": 0.013670518063008785, + "learning_rate": 0.0006, + "loss": 4.088400840759277, + "step": 6951 + }, + { + "epoch": 96.5589519650655, + "grad_norm": 0.013920796103775501, + "learning_rate": 0.0006, + "loss": 4.026619911193848, + "step": 6952 + }, + { + "epoch": 96.57292576419214, + "grad_norm": 0.013115637935698032, + "learning_rate": 0.0006, + "loss": 4.141090393066406, + "step": 6953 + }, + { + "epoch": 96.58689956331878, + "grad_norm": 0.013466455042362213, + "learning_rate": 0.0006, + "loss": 4.048084259033203, + "step": 6954 + }, + { + "epoch": 96.60087336244541, + "grad_norm": 0.013038012199103832, + "learning_rate": 0.0006, + "loss": 4.034649848937988, + "step": 6955 + }, + { + "epoch": 96.61484716157206, + "grad_norm": 0.012926512397825718, + "learning_rate": 0.0006, + "loss": 4.138445854187012, + "step": 6956 + }, + { + "epoch": 96.62882096069869, + "grad_norm": 0.013271111994981766, + "learning_rate": 0.0006, + "loss": 4.240598201751709, + "step": 6957 + }, + { + "epoch": 96.64279475982532, + "grad_norm": 0.013092855922877789, + "learning_rate": 0.0006, + "loss": 4.091870307922363, + "step": 6958 + }, + { + "epoch": 96.65676855895197, + "grad_norm": 0.013304130174219608, + "learning_rate": 0.0006, + "loss": 4.048708915710449, + "step": 6959 + }, + { + "epoch": 96.6707423580786, + "grad_norm": 0.01289574895054102, + "learning_rate": 0.0006, + "loss": 4.083846092224121, + "step": 6960 + }, + { + "epoch": 96.68471615720524, + "grad_norm": 0.012587146833539009, + "learning_rate": 0.0006, + "loss": 4.1233015060424805, + "step": 6961 + }, + { + "epoch": 96.69868995633188, + "grad_norm": 0.012312786653637886, + "learning_rate": 0.0006, + "loss": 4.0938262939453125, + "step": 6962 + }, + { + "epoch": 96.71266375545852, + "grad_norm": 0.014632516540586948, + "learning_rate": 0.0006, + "loss": 4.0226545333862305, + "step": 6963 + }, + { + "epoch": 96.72663755458515, + "grad_norm": 0.015110369771718979, + "learning_rate": 0.0006, + "loss": 4.052829742431641, + "step": 6964 + }, + { + "epoch": 96.74061135371178, + "grad_norm": 0.015997016802430153, + "learning_rate": 0.0006, + "loss": 4.126044750213623, + "step": 6965 + }, + { + "epoch": 96.75458515283843, + "grad_norm": 0.016428956761956215, + "learning_rate": 0.0006, + "loss": 4.126011848449707, + "step": 6966 + }, + { + "epoch": 96.76855895196506, + "grad_norm": 0.01638922467827797, + "learning_rate": 0.0006, + "loss": 4.1207685470581055, + "step": 6967 + }, + { + "epoch": 96.78253275109171, + "grad_norm": 0.01387663371860981, + "learning_rate": 0.0006, + "loss": 3.9403412342071533, + "step": 6968 + }, + { + "epoch": 96.79650655021834, + "grad_norm": 0.014120660722255707, + "learning_rate": 0.0006, + "loss": 4.202868461608887, + "step": 6969 + }, + { + "epoch": 96.81048034934499, + "grad_norm": 0.014854361303150654, + "learning_rate": 0.0006, + "loss": 4.015185832977295, + "step": 6970 + }, + { + "epoch": 96.82445414847162, + "grad_norm": 0.014359727501869202, + "learning_rate": 0.0006, + "loss": 4.119946479797363, + "step": 6971 + }, + { + "epoch": 96.83842794759825, + "grad_norm": 0.01564539223909378, + "learning_rate": 0.0006, + "loss": 3.9208037853240967, + "step": 6972 + }, + { + "epoch": 96.8524017467249, + "grad_norm": 0.015814676880836487, + "learning_rate": 0.0006, + "loss": 4.056029319763184, + "step": 6973 + }, + { + "epoch": 96.86637554585153, + "grad_norm": 0.016342822462320328, + "learning_rate": 0.0006, + "loss": 4.160576343536377, + "step": 6974 + }, + { + "epoch": 96.88034934497817, + "grad_norm": 0.016163306310772896, + "learning_rate": 0.0006, + "loss": 3.94950008392334, + "step": 6975 + }, + { + "epoch": 96.8943231441048, + "grad_norm": 0.018308300524950027, + "learning_rate": 0.0006, + "loss": 4.096719741821289, + "step": 6976 + }, + { + "epoch": 96.90829694323143, + "grad_norm": 0.01985356956720352, + "learning_rate": 0.0006, + "loss": 4.052276611328125, + "step": 6977 + }, + { + "epoch": 96.92227074235808, + "grad_norm": 0.01949184387922287, + "learning_rate": 0.0006, + "loss": 4.077932357788086, + "step": 6978 + }, + { + "epoch": 96.93624454148471, + "grad_norm": 0.018964678049087524, + "learning_rate": 0.0006, + "loss": 4.019469738006592, + "step": 6979 + }, + { + "epoch": 96.95021834061136, + "grad_norm": 0.019365238025784492, + "learning_rate": 0.0006, + "loss": 4.032814979553223, + "step": 6980 + }, + { + "epoch": 96.96419213973799, + "grad_norm": 0.01766037568449974, + "learning_rate": 0.0006, + "loss": 4.061336517333984, + "step": 6981 + }, + { + "epoch": 96.97816593886463, + "grad_norm": 0.018002403900027275, + "learning_rate": 0.0006, + "loss": 4.054131507873535, + "step": 6982 + }, + { + "epoch": 96.99213973799127, + "grad_norm": 0.01888604648411274, + "learning_rate": 0.0006, + "loss": 3.9295921325683594, + "step": 6983 + }, + { + "epoch": 97.0, + "grad_norm": 0.01609404757618904, + "learning_rate": 0.0006, + "loss": 4.037106513977051, + "step": 6984 + }, + { + "epoch": 97.0, + "eval_loss": 4.5986433029174805, + "eval_runtime": 56.5822, + "eval_samples_per_second": 43.158, + "eval_steps_per_second": 1.361, + "step": 6984 + }, + { + "epoch": 97.01397379912663, + "grad_norm": 0.015459981746971607, + "learning_rate": 0.0006, + "loss": 4.06088924407959, + "step": 6985 + }, + { + "epoch": 97.02794759825328, + "grad_norm": 0.016333693638443947, + "learning_rate": 0.0006, + "loss": 3.9525675773620605, + "step": 6986 + }, + { + "epoch": 97.04192139737991, + "grad_norm": 0.0157504603266716, + "learning_rate": 0.0006, + "loss": 4.143257141113281, + "step": 6987 + }, + { + "epoch": 97.05589519650655, + "grad_norm": 0.015252048149704933, + "learning_rate": 0.0006, + "loss": 4.158165454864502, + "step": 6988 + }, + { + "epoch": 97.06986899563319, + "grad_norm": 0.015041683800518513, + "learning_rate": 0.0006, + "loss": 4.030884742736816, + "step": 6989 + }, + { + "epoch": 97.08384279475983, + "grad_norm": 0.014693699777126312, + "learning_rate": 0.0006, + "loss": 3.8566339015960693, + "step": 6990 + }, + { + "epoch": 97.09781659388646, + "grad_norm": 0.0141681469976902, + "learning_rate": 0.0006, + "loss": 4.001909255981445, + "step": 6991 + }, + { + "epoch": 97.1117903930131, + "grad_norm": 0.013231917284429073, + "learning_rate": 0.0006, + "loss": 3.995041608810425, + "step": 6992 + }, + { + "epoch": 97.12576419213974, + "grad_norm": 0.014617119915783405, + "learning_rate": 0.0006, + "loss": 4.012053489685059, + "step": 6993 + }, + { + "epoch": 97.13973799126637, + "grad_norm": 0.014983531087636948, + "learning_rate": 0.0006, + "loss": 4.084125518798828, + "step": 6994 + }, + { + "epoch": 97.15371179039302, + "grad_norm": 0.013970102183520794, + "learning_rate": 0.0006, + "loss": 4.001402378082275, + "step": 6995 + }, + { + "epoch": 97.16768558951965, + "grad_norm": 0.0134900426492095, + "learning_rate": 0.0006, + "loss": 4.007713794708252, + "step": 6996 + }, + { + "epoch": 97.18165938864628, + "grad_norm": 0.013806418515741825, + "learning_rate": 0.0006, + "loss": 3.9524781703948975, + "step": 6997 + }, + { + "epoch": 97.19563318777293, + "grad_norm": 0.014304264448583126, + "learning_rate": 0.0006, + "loss": 3.9666526317596436, + "step": 6998 + }, + { + "epoch": 97.20960698689956, + "grad_norm": 0.016392510384321213, + "learning_rate": 0.0006, + "loss": 3.9915683269500732, + "step": 6999 + }, + { + "epoch": 97.2235807860262, + "grad_norm": 0.017044387757778168, + "learning_rate": 0.0006, + "loss": 4.11865234375, + "step": 7000 + }, + { + "epoch": 97.23755458515284, + "grad_norm": 0.016325082629919052, + "learning_rate": 0.0006, + "loss": 4.0636820793151855, + "step": 7001 + }, + { + "epoch": 97.25152838427948, + "grad_norm": 0.014634544961154461, + "learning_rate": 0.0006, + "loss": 3.967127799987793, + "step": 7002 + }, + { + "epoch": 97.26550218340611, + "grad_norm": 0.014043385162949562, + "learning_rate": 0.0006, + "loss": 4.072741985321045, + "step": 7003 + }, + { + "epoch": 97.27947598253274, + "grad_norm": 0.01376431342214346, + "learning_rate": 0.0006, + "loss": 4.07547664642334, + "step": 7004 + }, + { + "epoch": 97.29344978165939, + "grad_norm": 0.013959003612399101, + "learning_rate": 0.0006, + "loss": 3.9862523078918457, + "step": 7005 + }, + { + "epoch": 97.30742358078602, + "grad_norm": 0.01420508325099945, + "learning_rate": 0.0006, + "loss": 4.085540294647217, + "step": 7006 + }, + { + "epoch": 97.32139737991267, + "grad_norm": 0.01613876223564148, + "learning_rate": 0.0006, + "loss": 4.025545120239258, + "step": 7007 + }, + { + "epoch": 97.3353711790393, + "grad_norm": 0.017000367864966393, + "learning_rate": 0.0006, + "loss": 4.047660827636719, + "step": 7008 + }, + { + "epoch": 97.34934497816595, + "grad_norm": 0.01646680012345314, + "learning_rate": 0.0006, + "loss": 4.087928771972656, + "step": 7009 + }, + { + "epoch": 97.36331877729258, + "grad_norm": 0.015728263184428215, + "learning_rate": 0.0006, + "loss": 4.022225856781006, + "step": 7010 + }, + { + "epoch": 97.37729257641921, + "grad_norm": 0.015262200497090816, + "learning_rate": 0.0006, + "loss": 3.944697618484497, + "step": 7011 + }, + { + "epoch": 97.39126637554585, + "grad_norm": 0.016452079638838768, + "learning_rate": 0.0006, + "loss": 3.957644462585449, + "step": 7012 + }, + { + "epoch": 97.40524017467249, + "grad_norm": 0.018871447071433067, + "learning_rate": 0.0006, + "loss": 4.017784118652344, + "step": 7013 + }, + { + "epoch": 97.41921397379913, + "grad_norm": 0.01902632787823677, + "learning_rate": 0.0006, + "loss": 3.973320484161377, + "step": 7014 + }, + { + "epoch": 97.43318777292576, + "grad_norm": 0.016632359474897385, + "learning_rate": 0.0006, + "loss": 4.047779083251953, + "step": 7015 + }, + { + "epoch": 97.44716157205241, + "grad_norm": 0.01676122285425663, + "learning_rate": 0.0006, + "loss": 4.149074554443359, + "step": 7016 + }, + { + "epoch": 97.46113537117904, + "grad_norm": 0.01764233037829399, + "learning_rate": 0.0006, + "loss": 3.92502760887146, + "step": 7017 + }, + { + "epoch": 97.47510917030567, + "grad_norm": 0.01836075820028782, + "learning_rate": 0.0006, + "loss": 4.173796653747559, + "step": 7018 + }, + { + "epoch": 97.48908296943232, + "grad_norm": 0.017320366576313972, + "learning_rate": 0.0006, + "loss": 4.007115364074707, + "step": 7019 + }, + { + "epoch": 97.50305676855895, + "grad_norm": 0.015976276248693466, + "learning_rate": 0.0006, + "loss": 4.165112495422363, + "step": 7020 + }, + { + "epoch": 97.5170305676856, + "grad_norm": 0.0173528753221035, + "learning_rate": 0.0006, + "loss": 4.158927917480469, + "step": 7021 + }, + { + "epoch": 97.53100436681223, + "grad_norm": 0.01881110668182373, + "learning_rate": 0.0006, + "loss": 4.069684982299805, + "step": 7022 + }, + { + "epoch": 97.54497816593886, + "grad_norm": 0.01849045790731907, + "learning_rate": 0.0006, + "loss": 4.085758209228516, + "step": 7023 + }, + { + "epoch": 97.5589519650655, + "grad_norm": 0.016648925840854645, + "learning_rate": 0.0006, + "loss": 4.047055721282959, + "step": 7024 + }, + { + "epoch": 97.57292576419214, + "grad_norm": 0.014700878411531448, + "learning_rate": 0.0006, + "loss": 4.0177507400512695, + "step": 7025 + }, + { + "epoch": 97.58689956331878, + "grad_norm": 0.014601817354559898, + "learning_rate": 0.0006, + "loss": 4.070394039154053, + "step": 7026 + }, + { + "epoch": 97.60087336244541, + "grad_norm": 0.016391467303037643, + "learning_rate": 0.0006, + "loss": 3.957401990890503, + "step": 7027 + }, + { + "epoch": 97.61484716157206, + "grad_norm": 0.015401630662381649, + "learning_rate": 0.0006, + "loss": 4.040555000305176, + "step": 7028 + }, + { + "epoch": 97.62882096069869, + "grad_norm": 0.014783227816224098, + "learning_rate": 0.0006, + "loss": 4.039798259735107, + "step": 7029 + }, + { + "epoch": 97.64279475982532, + "grad_norm": 0.014704087749123573, + "learning_rate": 0.0006, + "loss": 4.059834003448486, + "step": 7030 + }, + { + "epoch": 97.65676855895197, + "grad_norm": 0.01375004556030035, + "learning_rate": 0.0006, + "loss": 4.084839344024658, + "step": 7031 + }, + { + "epoch": 97.6707423580786, + "grad_norm": 0.01403744611889124, + "learning_rate": 0.0006, + "loss": 4.060206890106201, + "step": 7032 + }, + { + "epoch": 97.68471615720524, + "grad_norm": 0.012827585451304913, + "learning_rate": 0.0006, + "loss": 4.187256813049316, + "step": 7033 + }, + { + "epoch": 97.69868995633188, + "grad_norm": 0.013072148896753788, + "learning_rate": 0.0006, + "loss": 4.136262893676758, + "step": 7034 + }, + { + "epoch": 97.71266375545852, + "grad_norm": 0.013496107421815395, + "learning_rate": 0.0006, + "loss": 3.9419479370117188, + "step": 7035 + }, + { + "epoch": 97.72663755458515, + "grad_norm": 0.014445106498897076, + "learning_rate": 0.0006, + "loss": 3.9494845867156982, + "step": 7036 + }, + { + "epoch": 97.74061135371178, + "grad_norm": 0.014507179148495197, + "learning_rate": 0.0006, + "loss": 4.05567741394043, + "step": 7037 + }, + { + "epoch": 97.75458515283843, + "grad_norm": 0.015422643162310123, + "learning_rate": 0.0006, + "loss": 4.103999137878418, + "step": 7038 + }, + { + "epoch": 97.76855895196506, + "grad_norm": 0.01515049859881401, + "learning_rate": 0.0006, + "loss": 4.094231605529785, + "step": 7039 + }, + { + "epoch": 97.78253275109171, + "grad_norm": 0.013850145973265171, + "learning_rate": 0.0006, + "loss": 4.084870338439941, + "step": 7040 + }, + { + "epoch": 97.79650655021834, + "grad_norm": 0.013332060538232327, + "learning_rate": 0.0006, + "loss": 4.193096160888672, + "step": 7041 + }, + { + "epoch": 97.81048034934499, + "grad_norm": 0.014778176322579384, + "learning_rate": 0.0006, + "loss": 4.0440216064453125, + "step": 7042 + }, + { + "epoch": 97.82445414847162, + "grad_norm": 0.013498589396476746, + "learning_rate": 0.0006, + "loss": 4.088800430297852, + "step": 7043 + }, + { + "epoch": 97.83842794759825, + "grad_norm": 0.014520619064569473, + "learning_rate": 0.0006, + "loss": 4.156583786010742, + "step": 7044 + }, + { + "epoch": 97.8524017467249, + "grad_norm": 0.014858689159154892, + "learning_rate": 0.0006, + "loss": 4.113174915313721, + "step": 7045 + }, + { + "epoch": 97.86637554585153, + "grad_norm": 0.015516536310315132, + "learning_rate": 0.0006, + "loss": 4.07958984375, + "step": 7046 + }, + { + "epoch": 97.88034934497817, + "grad_norm": 0.016619030386209488, + "learning_rate": 0.0006, + "loss": 4.03786563873291, + "step": 7047 + }, + { + "epoch": 97.8943231441048, + "grad_norm": 0.015789631754159927, + "learning_rate": 0.0006, + "loss": 4.065053462982178, + "step": 7048 + }, + { + "epoch": 97.90829694323143, + "grad_norm": 0.015848442912101746, + "learning_rate": 0.0006, + "loss": 4.183546543121338, + "step": 7049 + }, + { + "epoch": 97.92227074235808, + "grad_norm": 0.014621895737946033, + "learning_rate": 0.0006, + "loss": 4.024744987487793, + "step": 7050 + }, + { + "epoch": 97.93624454148471, + "grad_norm": 0.015793966129422188, + "learning_rate": 0.0006, + "loss": 3.9715819358825684, + "step": 7051 + }, + { + "epoch": 97.95021834061136, + "grad_norm": 0.014637742191553116, + "learning_rate": 0.0006, + "loss": 4.061223983764648, + "step": 7052 + }, + { + "epoch": 97.96419213973799, + "grad_norm": 0.0143470773473382, + "learning_rate": 0.0006, + "loss": 4.00929069519043, + "step": 7053 + }, + { + "epoch": 97.97816593886463, + "grad_norm": 0.014167622663080692, + "learning_rate": 0.0006, + "loss": 3.982048749923706, + "step": 7054 + }, + { + "epoch": 97.99213973799127, + "grad_norm": 0.01527823694050312, + "learning_rate": 0.0006, + "loss": 4.063419818878174, + "step": 7055 + }, + { + "epoch": 98.0, + "grad_norm": 0.015410988591611385, + "learning_rate": 0.0006, + "loss": 4.058195114135742, + "step": 7056 + }, + { + "epoch": 98.0, + "eval_loss": 4.5309038162231445, + "eval_runtime": 56.6931, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 7056 + }, + { + "epoch": 98.01397379912663, + "grad_norm": 0.013760345987975597, + "learning_rate": 0.0006, + "loss": 4.059067249298096, + "step": 7057 + }, + { + "epoch": 98.02794759825328, + "grad_norm": 0.013799279928207397, + "learning_rate": 0.0006, + "loss": 3.975419044494629, + "step": 7058 + }, + { + "epoch": 98.04192139737991, + "grad_norm": 0.014793423935770988, + "learning_rate": 0.0006, + "loss": 4.163051605224609, + "step": 7059 + }, + { + "epoch": 98.05589519650655, + "grad_norm": 0.015820559114217758, + "learning_rate": 0.0006, + "loss": 4.032471656799316, + "step": 7060 + }, + { + "epoch": 98.06986899563319, + "grad_norm": 0.014252204447984695, + "learning_rate": 0.0006, + "loss": 4.038850784301758, + "step": 7061 + }, + { + "epoch": 98.08384279475983, + "grad_norm": 0.01272337045520544, + "learning_rate": 0.0006, + "loss": 4.054554462432861, + "step": 7062 + }, + { + "epoch": 98.09781659388646, + "grad_norm": 0.01498495228588581, + "learning_rate": 0.0006, + "loss": 4.009291648864746, + "step": 7063 + }, + { + "epoch": 98.1117903930131, + "grad_norm": 0.015501809306442738, + "learning_rate": 0.0006, + "loss": 3.999412775039673, + "step": 7064 + }, + { + "epoch": 98.12576419213974, + "grad_norm": 0.015108918771147728, + "learning_rate": 0.0006, + "loss": 4.015168190002441, + "step": 7065 + }, + { + "epoch": 98.13973799126637, + "grad_norm": 0.01489692460745573, + "learning_rate": 0.0006, + "loss": 4.163906097412109, + "step": 7066 + }, + { + "epoch": 98.15371179039302, + "grad_norm": 0.014246871694922447, + "learning_rate": 0.0006, + "loss": 3.993173122406006, + "step": 7067 + }, + { + "epoch": 98.16768558951965, + "grad_norm": 0.015163703821599483, + "learning_rate": 0.0006, + "loss": 3.908207893371582, + "step": 7068 + }, + { + "epoch": 98.18165938864628, + "grad_norm": 0.0141439288854599, + "learning_rate": 0.0006, + "loss": 4.064990997314453, + "step": 7069 + }, + { + "epoch": 98.19563318777293, + "grad_norm": 0.013251710683107376, + "learning_rate": 0.0006, + "loss": 4.042545318603516, + "step": 7070 + }, + { + "epoch": 98.20960698689956, + "grad_norm": 0.01428674440830946, + "learning_rate": 0.0006, + "loss": 4.041232585906982, + "step": 7071 + }, + { + "epoch": 98.2235807860262, + "grad_norm": 0.012818179093301296, + "learning_rate": 0.0006, + "loss": 3.9160139560699463, + "step": 7072 + }, + { + "epoch": 98.23755458515284, + "grad_norm": 0.012890368700027466, + "learning_rate": 0.0006, + "loss": 3.9687235355377197, + "step": 7073 + }, + { + "epoch": 98.25152838427948, + "grad_norm": 0.014132961630821228, + "learning_rate": 0.0006, + "loss": 4.066800117492676, + "step": 7074 + }, + { + "epoch": 98.26550218340611, + "grad_norm": 0.013544383458793163, + "learning_rate": 0.0006, + "loss": 4.111777305603027, + "step": 7075 + }, + { + "epoch": 98.27947598253274, + "grad_norm": 0.014414685778319836, + "learning_rate": 0.0006, + "loss": 4.109633445739746, + "step": 7076 + }, + { + "epoch": 98.29344978165939, + "grad_norm": 0.015201127156615257, + "learning_rate": 0.0006, + "loss": 4.033087253570557, + "step": 7077 + }, + { + "epoch": 98.30742358078602, + "grad_norm": 0.016936076804995537, + "learning_rate": 0.0006, + "loss": 4.10897970199585, + "step": 7078 + }, + { + "epoch": 98.32139737991267, + "grad_norm": 0.01556481048464775, + "learning_rate": 0.0006, + "loss": 3.968942165374756, + "step": 7079 + }, + { + "epoch": 98.3353711790393, + "grad_norm": 0.015971971675753593, + "learning_rate": 0.0006, + "loss": 4.024805068969727, + "step": 7080 + }, + { + "epoch": 98.34934497816595, + "grad_norm": 0.016742313280701637, + "learning_rate": 0.0006, + "loss": 4.037862777709961, + "step": 7081 + }, + { + "epoch": 98.36331877729258, + "grad_norm": 0.015753688290715218, + "learning_rate": 0.0006, + "loss": 4.015701770782471, + "step": 7082 + }, + { + "epoch": 98.37729257641921, + "grad_norm": 0.014211287721991539, + "learning_rate": 0.0006, + "loss": 4.068171501159668, + "step": 7083 + }, + { + "epoch": 98.39126637554585, + "grad_norm": 0.013903875835239887, + "learning_rate": 0.0006, + "loss": 4.052516937255859, + "step": 7084 + }, + { + "epoch": 98.40524017467249, + "grad_norm": 0.015590954571962357, + "learning_rate": 0.0006, + "loss": 4.056125164031982, + "step": 7085 + }, + { + "epoch": 98.41921397379913, + "grad_norm": 0.016314011067152023, + "learning_rate": 0.0006, + "loss": 4.014211177825928, + "step": 7086 + }, + { + "epoch": 98.43318777292576, + "grad_norm": 0.017208732664585114, + "learning_rate": 0.0006, + "loss": 4.0615339279174805, + "step": 7087 + }, + { + "epoch": 98.44716157205241, + "grad_norm": 0.017213109880685806, + "learning_rate": 0.0006, + "loss": 3.9580445289611816, + "step": 7088 + }, + { + "epoch": 98.46113537117904, + "grad_norm": 0.016858434304594994, + "learning_rate": 0.0006, + "loss": 3.9932594299316406, + "step": 7089 + }, + { + "epoch": 98.47510917030567, + "grad_norm": 0.014381436631083488, + "learning_rate": 0.0006, + "loss": 4.159034729003906, + "step": 7090 + }, + { + "epoch": 98.48908296943232, + "grad_norm": 0.013713826425373554, + "learning_rate": 0.0006, + "loss": 4.055055141448975, + "step": 7091 + }, + { + "epoch": 98.50305676855895, + "grad_norm": 0.014764957129955292, + "learning_rate": 0.0006, + "loss": 4.059102535247803, + "step": 7092 + }, + { + "epoch": 98.5170305676856, + "grad_norm": 0.014522310346364975, + "learning_rate": 0.0006, + "loss": 4.093504905700684, + "step": 7093 + }, + { + "epoch": 98.53100436681223, + "grad_norm": 0.013484988361597061, + "learning_rate": 0.0006, + "loss": 4.079945087432861, + "step": 7094 + }, + { + "epoch": 98.54497816593886, + "grad_norm": 0.013231529854238033, + "learning_rate": 0.0006, + "loss": 4.046164512634277, + "step": 7095 + }, + { + "epoch": 98.5589519650655, + "grad_norm": 0.01392320916056633, + "learning_rate": 0.0006, + "loss": 3.975592613220215, + "step": 7096 + }, + { + "epoch": 98.57292576419214, + "grad_norm": 0.012323490343987942, + "learning_rate": 0.0006, + "loss": 3.9167721271514893, + "step": 7097 + }, + { + "epoch": 98.58689956331878, + "grad_norm": 0.012565829791128635, + "learning_rate": 0.0006, + "loss": 4.085109710693359, + "step": 7098 + }, + { + "epoch": 98.60087336244541, + "grad_norm": 0.013498359359800816, + "learning_rate": 0.0006, + "loss": 4.102974891662598, + "step": 7099 + }, + { + "epoch": 98.61484716157206, + "grad_norm": 0.014182298444211483, + "learning_rate": 0.0006, + "loss": 4.041601657867432, + "step": 7100 + }, + { + "epoch": 98.62882096069869, + "grad_norm": 0.013925517909228802, + "learning_rate": 0.0006, + "loss": 4.115788459777832, + "step": 7101 + }, + { + "epoch": 98.64279475982532, + "grad_norm": 0.03483536094427109, + "learning_rate": 0.0006, + "loss": 4.105683326721191, + "step": 7102 + }, + { + "epoch": 98.65676855895197, + "grad_norm": 0.014278477989137173, + "learning_rate": 0.0006, + "loss": 4.04194974899292, + "step": 7103 + }, + { + "epoch": 98.6707423580786, + "grad_norm": 0.015309838578104973, + "learning_rate": 0.0006, + "loss": 4.094045162200928, + "step": 7104 + }, + { + "epoch": 98.68471615720524, + "grad_norm": 0.015831857919692993, + "learning_rate": 0.0006, + "loss": 3.9989206790924072, + "step": 7105 + }, + { + "epoch": 98.69868995633188, + "grad_norm": 0.014071783982217312, + "learning_rate": 0.0006, + "loss": 4.0480852127075195, + "step": 7106 + }, + { + "epoch": 98.71266375545852, + "grad_norm": 0.014099884778261185, + "learning_rate": 0.0006, + "loss": 4.0519633293151855, + "step": 7107 + }, + { + "epoch": 98.72663755458515, + "grad_norm": 0.0143047496676445, + "learning_rate": 0.0006, + "loss": 3.9782276153564453, + "step": 7108 + }, + { + "epoch": 98.74061135371178, + "grad_norm": 0.013741862028837204, + "learning_rate": 0.0006, + "loss": 4.035103797912598, + "step": 7109 + }, + { + "epoch": 98.75458515283843, + "grad_norm": 0.014014648273587227, + "learning_rate": 0.0006, + "loss": 4.09281063079834, + "step": 7110 + }, + { + "epoch": 98.76855895196506, + "grad_norm": 0.013909446075558662, + "learning_rate": 0.0006, + "loss": 4.033784866333008, + "step": 7111 + }, + { + "epoch": 98.78253275109171, + "grad_norm": 0.013188562355935574, + "learning_rate": 0.0006, + "loss": 4.095016956329346, + "step": 7112 + }, + { + "epoch": 98.79650655021834, + "grad_norm": 0.013292612507939339, + "learning_rate": 0.0006, + "loss": 4.000730514526367, + "step": 7113 + }, + { + "epoch": 98.81048034934499, + "grad_norm": 0.013189040124416351, + "learning_rate": 0.0006, + "loss": 4.087883949279785, + "step": 7114 + }, + { + "epoch": 98.82445414847162, + "grad_norm": 0.013953711837530136, + "learning_rate": 0.0006, + "loss": 4.051388740539551, + "step": 7115 + }, + { + "epoch": 98.83842794759825, + "grad_norm": 0.014403220266103745, + "learning_rate": 0.0006, + "loss": 3.9786248207092285, + "step": 7116 + }, + { + "epoch": 98.8524017467249, + "grad_norm": 0.013747540302574635, + "learning_rate": 0.0006, + "loss": 3.9938254356384277, + "step": 7117 + }, + { + "epoch": 98.86637554585153, + "grad_norm": 0.014865634962916374, + "learning_rate": 0.0006, + "loss": 4.075680255889893, + "step": 7118 + }, + { + "epoch": 98.88034934497817, + "grad_norm": 0.015241672284901142, + "learning_rate": 0.0006, + "loss": 3.955733299255371, + "step": 7119 + }, + { + "epoch": 98.8943231441048, + "grad_norm": 0.013307510875165462, + "learning_rate": 0.0006, + "loss": 3.94956374168396, + "step": 7120 + }, + { + "epoch": 98.90829694323143, + "grad_norm": 0.013500694185495377, + "learning_rate": 0.0006, + "loss": 4.120021343231201, + "step": 7121 + }, + { + "epoch": 98.92227074235808, + "grad_norm": 0.015430806204676628, + "learning_rate": 0.0006, + "loss": 4.108362674713135, + "step": 7122 + }, + { + "epoch": 98.93624454148471, + "grad_norm": 0.015447405166924, + "learning_rate": 0.0006, + "loss": 4.040914058685303, + "step": 7123 + }, + { + "epoch": 98.95021834061136, + "grad_norm": 0.014040131121873856, + "learning_rate": 0.0006, + "loss": 4.086688995361328, + "step": 7124 + }, + { + "epoch": 98.96419213973799, + "grad_norm": 0.0136969443410635, + "learning_rate": 0.0006, + "loss": 4.083828926086426, + "step": 7125 + }, + { + "epoch": 98.97816593886463, + "grad_norm": 0.01413224171847105, + "learning_rate": 0.0006, + "loss": 4.142146110534668, + "step": 7126 + }, + { + "epoch": 98.99213973799127, + "grad_norm": 0.012932452373206615, + "learning_rate": 0.0006, + "loss": 4.1193342208862305, + "step": 7127 + }, + { + "epoch": 99.0, + "grad_norm": 0.01586783304810524, + "learning_rate": 0.0006, + "loss": 4.191825866699219, + "step": 7128 + }, + { + "epoch": 99.0, + "eval_loss": 4.550753116607666, + "eval_runtime": 56.4315, + "eval_samples_per_second": 43.274, + "eval_steps_per_second": 1.364, + "step": 7128 + }, + { + "epoch": 99.01397379912663, + "grad_norm": 0.016847660765051842, + "learning_rate": 0.0006, + "loss": 3.9874870777130127, + "step": 7129 + }, + { + "epoch": 99.02794759825328, + "grad_norm": 0.016426295042037964, + "learning_rate": 0.0006, + "loss": 4.108767032623291, + "step": 7130 + }, + { + "epoch": 99.04192139737991, + "grad_norm": 0.01687243953347206, + "learning_rate": 0.0006, + "loss": 3.9658308029174805, + "step": 7131 + }, + { + "epoch": 99.05589519650655, + "grad_norm": 0.016468465328216553, + "learning_rate": 0.0006, + "loss": 3.91381573677063, + "step": 7132 + }, + { + "epoch": 99.06986899563319, + "grad_norm": 0.013777337968349457, + "learning_rate": 0.0006, + "loss": 3.9960741996765137, + "step": 7133 + }, + { + "epoch": 99.08384279475983, + "grad_norm": 0.014061779715120792, + "learning_rate": 0.0006, + "loss": 4.069479942321777, + "step": 7134 + }, + { + "epoch": 99.09781659388646, + "grad_norm": 0.01513410173356533, + "learning_rate": 0.0006, + "loss": 4.145164489746094, + "step": 7135 + }, + { + "epoch": 99.1117903930131, + "grad_norm": 0.01567850261926651, + "learning_rate": 0.0006, + "loss": 4.016868591308594, + "step": 7136 + }, + { + "epoch": 99.12576419213974, + "grad_norm": 0.01662704534828663, + "learning_rate": 0.0006, + "loss": 3.966663360595703, + "step": 7137 + }, + { + "epoch": 99.13973799126637, + "grad_norm": 0.01690712384879589, + "learning_rate": 0.0006, + "loss": 4.003656387329102, + "step": 7138 + }, + { + "epoch": 99.15371179039302, + "grad_norm": 0.01541030965745449, + "learning_rate": 0.0006, + "loss": 3.9990856647491455, + "step": 7139 + }, + { + "epoch": 99.16768558951965, + "grad_norm": 0.014438044279813766, + "learning_rate": 0.0006, + "loss": 4.089632034301758, + "step": 7140 + }, + { + "epoch": 99.18165938864628, + "grad_norm": 0.01470417995005846, + "learning_rate": 0.0006, + "loss": 4.018681049346924, + "step": 7141 + }, + { + "epoch": 99.19563318777293, + "grad_norm": 0.015567616559565067, + "learning_rate": 0.0006, + "loss": 4.067337989807129, + "step": 7142 + }, + { + "epoch": 99.20960698689956, + "grad_norm": 0.01784040965139866, + "learning_rate": 0.0006, + "loss": 4.083293914794922, + "step": 7143 + }, + { + "epoch": 99.2235807860262, + "grad_norm": 0.017351210117340088, + "learning_rate": 0.0006, + "loss": 4.115224361419678, + "step": 7144 + }, + { + "epoch": 99.23755458515284, + "grad_norm": 0.014561209827661514, + "learning_rate": 0.0006, + "loss": 4.1001057624816895, + "step": 7145 + }, + { + "epoch": 99.25152838427948, + "grad_norm": 0.013986383564770222, + "learning_rate": 0.0006, + "loss": 3.9758031368255615, + "step": 7146 + }, + { + "epoch": 99.26550218340611, + "grad_norm": 0.014053978025913239, + "learning_rate": 0.0006, + "loss": 4.1420369148254395, + "step": 7147 + }, + { + "epoch": 99.27947598253274, + "grad_norm": 0.014878278598189354, + "learning_rate": 0.0006, + "loss": 4.018189430236816, + "step": 7148 + }, + { + "epoch": 99.29344978165939, + "grad_norm": 0.014634845778346062, + "learning_rate": 0.0006, + "loss": 4.062045574188232, + "step": 7149 + }, + { + "epoch": 99.30742358078602, + "grad_norm": 0.014713970944285393, + "learning_rate": 0.0006, + "loss": 4.017163276672363, + "step": 7150 + }, + { + "epoch": 99.32139737991267, + "grad_norm": 0.01589985005557537, + "learning_rate": 0.0006, + "loss": 4.093014717102051, + "step": 7151 + }, + { + "epoch": 99.3353711790393, + "grad_norm": 0.016232378780841827, + "learning_rate": 0.0006, + "loss": 4.029753684997559, + "step": 7152 + }, + { + "epoch": 99.34934497816595, + "grad_norm": 0.016340898349881172, + "learning_rate": 0.0006, + "loss": 4.056434631347656, + "step": 7153 + }, + { + "epoch": 99.36331877729258, + "grad_norm": 0.01587427407503128, + "learning_rate": 0.0006, + "loss": 3.932020425796509, + "step": 7154 + }, + { + "epoch": 99.37729257641921, + "grad_norm": 0.014977093786001205, + "learning_rate": 0.0006, + "loss": 4.015125751495361, + "step": 7155 + }, + { + "epoch": 99.39126637554585, + "grad_norm": 0.014813411049544811, + "learning_rate": 0.0006, + "loss": 4.088035583496094, + "step": 7156 + }, + { + "epoch": 99.40524017467249, + "grad_norm": 0.014407774433493614, + "learning_rate": 0.0006, + "loss": 3.9979054927825928, + "step": 7157 + }, + { + "epoch": 99.41921397379913, + "grad_norm": 0.013969994150102139, + "learning_rate": 0.0006, + "loss": 4.180194854736328, + "step": 7158 + }, + { + "epoch": 99.43318777292576, + "grad_norm": 0.0159455519169569, + "learning_rate": 0.0006, + "loss": 4.049341201782227, + "step": 7159 + }, + { + "epoch": 99.44716157205241, + "grad_norm": 0.016426580026745796, + "learning_rate": 0.0006, + "loss": 4.01624059677124, + "step": 7160 + }, + { + "epoch": 99.46113537117904, + "grad_norm": 0.014921349473297596, + "learning_rate": 0.0006, + "loss": 4.003138542175293, + "step": 7161 + }, + { + "epoch": 99.47510917030567, + "grad_norm": 0.014607159420847893, + "learning_rate": 0.0006, + "loss": 3.993896484375, + "step": 7162 + }, + { + "epoch": 99.48908296943232, + "grad_norm": 0.014638705179095268, + "learning_rate": 0.0006, + "loss": 4.015478134155273, + "step": 7163 + }, + { + "epoch": 99.50305676855895, + "grad_norm": 0.01346661988645792, + "learning_rate": 0.0006, + "loss": 3.9999020099639893, + "step": 7164 + }, + { + "epoch": 99.5170305676856, + "grad_norm": 0.013515726663172245, + "learning_rate": 0.0006, + "loss": 4.045872688293457, + "step": 7165 + }, + { + "epoch": 99.53100436681223, + "grad_norm": 0.013773965649306774, + "learning_rate": 0.0006, + "loss": 4.094198703765869, + "step": 7166 + }, + { + "epoch": 99.54497816593886, + "grad_norm": 0.014282289892435074, + "learning_rate": 0.0006, + "loss": 4.091470718383789, + "step": 7167 + }, + { + "epoch": 99.5589519650655, + "grad_norm": 0.015272347256541252, + "learning_rate": 0.0006, + "loss": 4.026535987854004, + "step": 7168 + }, + { + "epoch": 99.57292576419214, + "grad_norm": 0.015618950128555298, + "learning_rate": 0.0006, + "loss": 4.006448745727539, + "step": 7169 + }, + { + "epoch": 99.58689956331878, + "grad_norm": 0.015366379171609879, + "learning_rate": 0.0006, + "loss": 4.025970458984375, + "step": 7170 + }, + { + "epoch": 99.60087336244541, + "grad_norm": 0.013261756859719753, + "learning_rate": 0.0006, + "loss": 3.980401039123535, + "step": 7171 + }, + { + "epoch": 99.61484716157206, + "grad_norm": 0.013438165187835693, + "learning_rate": 0.0006, + "loss": 4.0106096267700195, + "step": 7172 + }, + { + "epoch": 99.62882096069869, + "grad_norm": 0.015591824427247047, + "learning_rate": 0.0006, + "loss": 4.0458784103393555, + "step": 7173 + }, + { + "epoch": 99.64279475982532, + "grad_norm": 0.014863507822155952, + "learning_rate": 0.0006, + "loss": 4.014740943908691, + "step": 7174 + }, + { + "epoch": 99.65676855895197, + "grad_norm": 0.012508790008723736, + "learning_rate": 0.0006, + "loss": 4.022995948791504, + "step": 7175 + }, + { + "epoch": 99.6707423580786, + "grad_norm": 0.013107656501233578, + "learning_rate": 0.0006, + "loss": 4.085544109344482, + "step": 7176 + }, + { + "epoch": 99.68471615720524, + "grad_norm": 0.013550374656915665, + "learning_rate": 0.0006, + "loss": 4.1239237785339355, + "step": 7177 + }, + { + "epoch": 99.69868995633188, + "grad_norm": 0.012947923503816128, + "learning_rate": 0.0006, + "loss": 4.01287317276001, + "step": 7178 + }, + { + "epoch": 99.71266375545852, + "grad_norm": 0.015236291103065014, + "learning_rate": 0.0006, + "loss": 4.075140476226807, + "step": 7179 + }, + { + "epoch": 99.72663755458515, + "grad_norm": 0.014862542040646076, + "learning_rate": 0.0006, + "loss": 4.111439228057861, + "step": 7180 + }, + { + "epoch": 99.74061135371178, + "grad_norm": 0.013883531093597412, + "learning_rate": 0.0006, + "loss": 3.9860856533050537, + "step": 7181 + }, + { + "epoch": 99.75458515283843, + "grad_norm": 0.01507870014756918, + "learning_rate": 0.0006, + "loss": 4.096504211425781, + "step": 7182 + }, + { + "epoch": 99.76855895196506, + "grad_norm": 0.015378627926111221, + "learning_rate": 0.0006, + "loss": 4.034113883972168, + "step": 7183 + }, + { + "epoch": 99.78253275109171, + "grad_norm": 0.013696327805519104, + "learning_rate": 0.0006, + "loss": 4.026954650878906, + "step": 7184 + }, + { + "epoch": 99.79650655021834, + "grad_norm": 0.013519738800823689, + "learning_rate": 0.0006, + "loss": 4.040035247802734, + "step": 7185 + }, + { + "epoch": 99.81048034934499, + "grad_norm": 0.015677500516176224, + "learning_rate": 0.0006, + "loss": 4.057739734649658, + "step": 7186 + }, + { + "epoch": 99.82445414847162, + "grad_norm": 0.014739335514605045, + "learning_rate": 0.0006, + "loss": 4.022522449493408, + "step": 7187 + }, + { + "epoch": 99.83842794759825, + "grad_norm": 0.014079970307648182, + "learning_rate": 0.0006, + "loss": 3.9286317825317383, + "step": 7188 + }, + { + "epoch": 99.8524017467249, + "grad_norm": 0.014899874106049538, + "learning_rate": 0.0006, + "loss": 4.081925868988037, + "step": 7189 + }, + { + "epoch": 99.86637554585153, + "grad_norm": 0.014767617918550968, + "learning_rate": 0.0006, + "loss": 4.029382705688477, + "step": 7190 + }, + { + "epoch": 99.88034934497817, + "grad_norm": 0.014914393424987793, + "learning_rate": 0.0006, + "loss": 4.056460857391357, + "step": 7191 + }, + { + "epoch": 99.8943231441048, + "grad_norm": 0.01720450073480606, + "learning_rate": 0.0006, + "loss": 4.0184478759765625, + "step": 7192 + }, + { + "epoch": 99.90829694323143, + "grad_norm": 0.016545087099075317, + "learning_rate": 0.0006, + "loss": 3.9638831615448, + "step": 7193 + }, + { + "epoch": 99.92227074235808, + "grad_norm": 0.014489716850221157, + "learning_rate": 0.0006, + "loss": 4.0267653465271, + "step": 7194 + }, + { + "epoch": 99.93624454148471, + "grad_norm": 0.01438206061720848, + "learning_rate": 0.0006, + "loss": 4.212850570678711, + "step": 7195 + }, + { + "epoch": 99.95021834061136, + "grad_norm": 0.017636865377426147, + "learning_rate": 0.0006, + "loss": 3.942308187484741, + "step": 7196 + }, + { + "epoch": 99.96419213973799, + "grad_norm": 0.019075388088822365, + "learning_rate": 0.0006, + "loss": 4.026660919189453, + "step": 7197 + }, + { + "epoch": 99.97816593886463, + "grad_norm": 0.016569023951888084, + "learning_rate": 0.0006, + "loss": 4.018840789794922, + "step": 7198 + }, + { + "epoch": 99.99213973799127, + "grad_norm": 0.014842798933386803, + "learning_rate": 0.0006, + "loss": 4.041904449462891, + "step": 7199 + }, + { + "epoch": 100.0, + "grad_norm": 0.0167684406042099, + "learning_rate": 0.0006, + "loss": 3.9079346656799316, + "step": 7200 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.069613361332224e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7200/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-7200/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7200/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-7488/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-7488/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f144376bd538aaf7ae17f67f42915e51e25e1606 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3ec22916dffc70ec6616d3654e31d00a4a10c6ca1d1ebb5120714499328f0e1 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7488/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5fbee4a581363ebb84c0136b3b253f02e01f4a89 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03d1a157ba5bd43a017408e321b80745d9c02dbb41a6dba2efb76d6c945344aa +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..61c0b7ef05f2f915a817eff75c694be3671b38f8 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46f7bd2ee110b8044cafb2a9f1de5db77692de8efefbdc19f5143ad386e6245d +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..fb2ee3987752f036205b24ae6005def818f4545a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d2ee4a11604a3be729b14d3e61f4e3d57002cafd565766464aec5477f1200da +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..e6c968a3d5ae020d0a1140c25257d1455b0fc9da --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ebce020dff1b60aa525ffeb3690e9655bea876e28d9e589bb23a37836707c3d +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..f1667e253147f06eace35355fed65f57719cd3f2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2b882c43396f99a07747905c96e2976f1484c5177ccc0805b42fcfc137980d0 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-7488/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..132f9886976cefcb2664879521dcc5aa82c41d3f --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af32b492fedaa58d69ad948e745403317a9384ef5bb7e3fec0e073cc81d818a2 +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-7488/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..69f2093714aae16a5aa9d2abe0d7bdb2cd3bd779 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/trainer_state.json @@ -0,0 +1,53274 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 104.0, + "eval_steps": 500, + "global_step": 7488, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.044651985168457, + "eval_runtime": 56.9145, + "eval_samples_per_second": 42.906, + "eval_steps_per_second": 1.353, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.017597166821360588, + "learning_rate": 0.0006, + "loss": 4.830524444580078, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.01716788113117218, + "learning_rate": 0.0006, + "loss": 4.894325256347656, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.017712457105517387, + "learning_rate": 0.0006, + "loss": 4.868297576904297, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.021486390382051468, + "learning_rate": 0.0006, + "loss": 4.922049522399902, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.024560654535889626, + "learning_rate": 0.0006, + "loss": 4.734241008758545, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.025870252400636673, + "learning_rate": 0.0006, + "loss": 4.868566989898682, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.0277389045804739, + "learning_rate": 0.0006, + "loss": 4.755006790161133, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.029100844636559486, + "learning_rate": 0.0006, + "loss": 4.909667015075684, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.031833067536354065, + "learning_rate": 0.0006, + "loss": 4.858674049377441, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.032709237188100815, + "learning_rate": 0.0006, + "loss": 4.823696136474609, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.02995380386710167, + "learning_rate": 0.0006, + "loss": 4.850235462188721, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.034098610281944275, + "learning_rate": 0.0006, + "loss": 4.8169989585876465, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.03377080336213112, + "learning_rate": 0.0006, + "loss": 4.90738582611084, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.029780825600028038, + "learning_rate": 0.0006, + "loss": 4.811163902282715, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.03118024580180645, + "learning_rate": 0.0006, + "loss": 4.861286640167236, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.033052004873752594, + "learning_rate": 0.0006, + "loss": 4.865021228790283, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.029070032760500908, + "learning_rate": 0.0006, + "loss": 4.947303771972656, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.028355715796351433, + "learning_rate": 0.0006, + "loss": 4.804648399353027, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.031257227063179016, + "learning_rate": 0.0006, + "loss": 4.757718086242676, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.031438373029232025, + "learning_rate": 0.0006, + "loss": 4.851268768310547, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.030492138117551804, + "learning_rate": 0.0006, + "loss": 4.7891035079956055, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026115991175174713, + "learning_rate": 0.0006, + "loss": 4.86474609375, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.024184465408325195, + "learning_rate": 0.0006, + "loss": 4.860456943511963, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025437982752919197, + "learning_rate": 0.0006, + "loss": 4.806251049041748, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.022390231490135193, + "learning_rate": 0.0006, + "loss": 4.807086944580078, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.020174294710159302, + "learning_rate": 0.0006, + "loss": 4.78292179107666, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.019734324887394905, + "learning_rate": 0.0006, + "loss": 4.784509658813477, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.016733523458242416, + "learning_rate": 0.0006, + "loss": 4.811755180358887, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.01570146717131138, + "learning_rate": 0.0006, + "loss": 4.720088005065918, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01601393334567547, + "learning_rate": 0.0006, + "loss": 4.783543109893799, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.014035487547516823, + "learning_rate": 0.0006, + "loss": 4.848340034484863, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015248487703502178, + "learning_rate": 0.0006, + "loss": 4.773287773132324, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01532268151640892, + "learning_rate": 0.0006, + "loss": 4.799646854400635, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.013982534408569336, + "learning_rate": 0.0006, + "loss": 4.8090500831604, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.014081164263188839, + "learning_rate": 0.0006, + "loss": 4.751888751983643, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01380133256316185, + "learning_rate": 0.0006, + "loss": 4.78144645690918, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013555269688367844, + "learning_rate": 0.0006, + "loss": 4.779257774353027, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.013075701892375946, + "learning_rate": 0.0006, + "loss": 4.820705413818359, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01276316400617361, + "learning_rate": 0.0006, + "loss": 4.836733818054199, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.013033892959356308, + "learning_rate": 0.0006, + "loss": 4.764178276062012, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.013996980153024197, + "learning_rate": 0.0006, + "loss": 4.767983436584473, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.015222184360027313, + "learning_rate": 0.0006, + "loss": 4.759721755981445, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.01623941957950592, + "learning_rate": 0.0006, + "loss": 4.849811553955078, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.015711182728409767, + "learning_rate": 0.0006, + "loss": 4.7829084396362305, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.014137467369437218, + "learning_rate": 0.0006, + "loss": 4.73321533203125, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.012738462537527084, + "learning_rate": 0.0006, + "loss": 4.870052337646484, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.013683130033314228, + "learning_rate": 0.0006, + "loss": 4.82658052444458, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014409830793738365, + "learning_rate": 0.0006, + "loss": 4.707494735717773, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.013806809671223164, + "learning_rate": 0.0006, + "loss": 4.778928279876709, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.012857912108302116, + "learning_rate": 0.0006, + "loss": 4.890717506408691, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.013828758150339127, + "learning_rate": 0.0006, + "loss": 4.8357439041137695, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01478351280093193, + "learning_rate": 0.0006, + "loss": 4.869986534118652, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.01326421182602644, + "learning_rate": 0.0006, + "loss": 4.830646514892578, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.013335433788597584, + "learning_rate": 0.0006, + "loss": 4.7786149978637695, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.014261147007346153, + "learning_rate": 0.0006, + "loss": 4.788693428039551, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.014274783432483673, + "learning_rate": 0.0006, + "loss": 4.77295446395874, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.015074674971401691, + "learning_rate": 0.0006, + "loss": 4.830754280090332, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.015599515289068222, + "learning_rate": 0.0006, + "loss": 4.837759017944336, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.014595242217183113, + "learning_rate": 0.0006, + "loss": 4.882965564727783, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.014951467514038086, + "learning_rate": 0.0006, + "loss": 4.709748268127441, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.01539295818656683, + "learning_rate": 0.0006, + "loss": 4.791160583496094, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015987906605005264, + "learning_rate": 0.0006, + "loss": 4.661876678466797, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.016527341678738594, + "learning_rate": 0.0006, + "loss": 4.721952438354492, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.01637161523103714, + "learning_rate": 0.0006, + "loss": 4.704446792602539, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.015221050009131432, + "learning_rate": 0.0006, + "loss": 4.771907806396484, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.016242166981101036, + "learning_rate": 0.0006, + "loss": 4.806267738342285, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.01628357358276844, + "learning_rate": 0.0006, + "loss": 4.7252326011657715, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.017428133636713028, + "learning_rate": 0.0006, + "loss": 4.765689849853516, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.01870061084628105, + "learning_rate": 0.0006, + "loss": 4.814488887786865, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.019796060398221016, + "learning_rate": 0.0006, + "loss": 4.739185810089111, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.01985820196568966, + "learning_rate": 0.0006, + "loss": 4.810608863830566, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.020834477618336678, + "learning_rate": 0.0006, + "loss": 4.727858066558838, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 4.979116916656494, + "eval_runtime": 57.2877, + "eval_samples_per_second": 42.627, + "eval_steps_per_second": 1.344, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.02025519125163555, + "learning_rate": 0.0006, + "loss": 4.628875732421875, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.019596368074417114, + "learning_rate": 0.0006, + "loss": 4.694699287414551, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.020494934171438217, + "learning_rate": 0.0006, + "loss": 4.691821098327637, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.020121760666370392, + "learning_rate": 0.0006, + "loss": 4.71478271484375, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.021091405302286148, + "learning_rate": 0.0006, + "loss": 4.695915222167969, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.019787423312664032, + "learning_rate": 0.0006, + "loss": 4.8652448654174805, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.0185833927243948, + "learning_rate": 0.0006, + "loss": 4.7065887451171875, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.021220475435256958, + "learning_rate": 0.0006, + "loss": 4.741232395172119, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.02136622555553913, + "learning_rate": 0.0006, + "loss": 4.705967903137207, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.0191587433218956, + "learning_rate": 0.0006, + "loss": 4.758667945861816, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.018290022388100624, + "learning_rate": 0.0006, + "loss": 4.760666847229004, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.021704409271478653, + "learning_rate": 0.0006, + "loss": 4.750555992126465, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.021668918430805206, + "learning_rate": 0.0006, + "loss": 4.840903282165527, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.019497141242027283, + "learning_rate": 0.0006, + "loss": 4.738969326019287, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.018500596284866333, + "learning_rate": 0.0006, + "loss": 4.75154972076416, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.01872239261865616, + "learning_rate": 0.0006, + "loss": 4.811738014221191, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.020321547985076904, + "learning_rate": 0.0006, + "loss": 4.815384387969971, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.02283575013279915, + "learning_rate": 0.0006, + "loss": 4.74582052230835, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.02474398910999298, + "learning_rate": 0.0006, + "loss": 4.738375186920166, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.02065764181315899, + "learning_rate": 0.0006, + "loss": 4.746875286102295, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.020534520968794823, + "learning_rate": 0.0006, + "loss": 4.656914710998535, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.0206378772854805, + "learning_rate": 0.0006, + "loss": 4.842615604400635, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.017935654148459435, + "learning_rate": 0.0006, + "loss": 4.717827796936035, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.016958363354206085, + "learning_rate": 0.0006, + "loss": 4.681070327758789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.015735412016510963, + "learning_rate": 0.0006, + "loss": 4.7957048416137695, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.01563699170947075, + "learning_rate": 0.0006, + "loss": 4.701192378997803, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.015207760035991669, + "learning_rate": 0.0006, + "loss": 4.73256254196167, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.015396883711218834, + "learning_rate": 0.0006, + "loss": 4.76412296295166, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.015690583735704422, + "learning_rate": 0.0006, + "loss": 4.791157245635986, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01755724661052227, + "learning_rate": 0.0006, + "loss": 4.845666885375977, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.019443422555923462, + "learning_rate": 0.0006, + "loss": 4.5902299880981445, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.02118338644504547, + "learning_rate": 0.0006, + "loss": 4.740627288818359, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.02174839936196804, + "learning_rate": 0.0006, + "loss": 4.814560890197754, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.022860374301671982, + "learning_rate": 0.0006, + "loss": 4.66440486907959, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.024742677807807922, + "learning_rate": 0.0006, + "loss": 4.7451653480529785, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.027068907395005226, + "learning_rate": 0.0006, + "loss": 4.802584171295166, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.026316625997424126, + "learning_rate": 0.0006, + "loss": 4.842397689819336, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.024610962718725204, + "learning_rate": 0.0006, + "loss": 4.650721073150635, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.024122603237628937, + "learning_rate": 0.0006, + "loss": 4.6984357833862305, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.022699708119034767, + "learning_rate": 0.0006, + "loss": 4.793007850646973, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.021706534549593925, + "learning_rate": 0.0006, + "loss": 4.691362380981445, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.021363291889429092, + "learning_rate": 0.0006, + "loss": 4.76762580871582, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.023476677015423775, + "learning_rate": 0.0006, + "loss": 4.777202606201172, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02457100711762905, + "learning_rate": 0.0006, + "loss": 4.590703010559082, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.02469099499285221, + "learning_rate": 0.0006, + "loss": 4.640185356140137, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.024914603680372238, + "learning_rate": 0.0006, + "loss": 4.772201061248779, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.025133173912763596, + "learning_rate": 0.0006, + "loss": 4.743337154388428, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.02034137211740017, + "learning_rate": 0.0006, + "loss": 4.752127647399902, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.019645661115646362, + "learning_rate": 0.0006, + "loss": 4.838453769683838, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.01900297962129116, + "learning_rate": 0.0006, + "loss": 4.804868221282959, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.020292341709136963, + "learning_rate": 0.0006, + "loss": 4.609743118286133, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.017014775425195694, + "learning_rate": 0.0006, + "loss": 4.740612983703613, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.015199844725430012, + "learning_rate": 0.0006, + "loss": 4.719701766967773, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.015125414356589317, + "learning_rate": 0.0006, + "loss": 4.686114311218262, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.014779540710151196, + "learning_rate": 0.0006, + "loss": 4.674347877502441, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.015255163423717022, + "learning_rate": 0.0006, + "loss": 4.6309309005737305, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.013692197389900684, + "learning_rate": 0.0006, + "loss": 4.698497295379639, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.014529623091220856, + "learning_rate": 0.0006, + "loss": 4.628443717956543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.01576659083366394, + "learning_rate": 0.0006, + "loss": 4.788015842437744, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.015834277495741844, + "learning_rate": 0.0006, + "loss": 4.672703742980957, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.015539503656327724, + "learning_rate": 0.0006, + "loss": 4.731503486633301, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013463746756315231, + "learning_rate": 0.0006, + "loss": 4.7006330490112305, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.016377726569771767, + "learning_rate": 0.0006, + "loss": 4.636074066162109, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.016660034656524658, + "learning_rate": 0.0006, + "loss": 4.742693901062012, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.014902938157320023, + "learning_rate": 0.0006, + "loss": 4.769811630249023, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.014883476309478283, + "learning_rate": 0.0006, + "loss": 4.731159210205078, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.014494217932224274, + "learning_rate": 0.0006, + "loss": 4.659485816955566, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.015426991507411003, + "learning_rate": 0.0006, + "loss": 4.755229949951172, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.016648586839437485, + "learning_rate": 0.0006, + "loss": 4.74111270904541, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.016980541869997978, + "learning_rate": 0.0006, + "loss": 4.74072265625, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.018070152029395103, + "learning_rate": 0.0006, + "loss": 4.732975006103516, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.019910665228962898, + "learning_rate": 0.0006, + "loss": 4.634697914123535, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 4.850640296936035, + "eval_runtime": 57.3039, + "eval_samples_per_second": 42.615, + "eval_steps_per_second": 1.344, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.01980053074657917, + "learning_rate": 0.0006, + "loss": 4.628744602203369, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02497623674571514, + "learning_rate": 0.0006, + "loss": 4.75930118560791, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.02794097363948822, + "learning_rate": 0.0006, + "loss": 4.691340446472168, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.023725535720586777, + "learning_rate": 0.0006, + "loss": 4.733664512634277, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.022966474294662476, + "learning_rate": 0.0006, + "loss": 4.664808750152588, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.02530563250184059, + "learning_rate": 0.0006, + "loss": 4.8047285079956055, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.02491082064807415, + "learning_rate": 0.0006, + "loss": 4.667956352233887, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.022110294550657272, + "learning_rate": 0.0006, + "loss": 4.666140556335449, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.02065849117934704, + "learning_rate": 0.0006, + "loss": 4.581960678100586, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.01941298507153988, + "learning_rate": 0.0006, + "loss": 4.614226341247559, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.022441856563091278, + "learning_rate": 0.0006, + "loss": 4.638352394104004, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.021519694477319717, + "learning_rate": 0.0006, + "loss": 4.700160980224609, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018956486135721207, + "learning_rate": 0.0006, + "loss": 4.649593353271484, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.016681140288710594, + "learning_rate": 0.0006, + "loss": 4.603952407836914, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.01788809522986412, + "learning_rate": 0.0006, + "loss": 4.669160842895508, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.016047228127717972, + "learning_rate": 0.0006, + "loss": 4.6367950439453125, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.01614239625632763, + "learning_rate": 0.0006, + "loss": 4.664529323577881, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.016362549737095833, + "learning_rate": 0.0006, + "loss": 4.702373504638672, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.015800409018993378, + "learning_rate": 0.0006, + "loss": 4.7594428062438965, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.016730139032006264, + "learning_rate": 0.0006, + "loss": 4.677244186401367, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.01661759801208973, + "learning_rate": 0.0006, + "loss": 4.633679389953613, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.017673347145318985, + "learning_rate": 0.0006, + "loss": 4.649107456207275, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.019243352115154266, + "learning_rate": 0.0006, + "loss": 4.611257553100586, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.019817164167761803, + "learning_rate": 0.0006, + "loss": 4.588086128234863, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.022292211651802063, + "learning_rate": 0.0006, + "loss": 4.783066749572754, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.02140096202492714, + "learning_rate": 0.0006, + "loss": 4.604718208312988, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.0187817495316267, + "learning_rate": 0.0006, + "loss": 4.721773147583008, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.019879210740327835, + "learning_rate": 0.0006, + "loss": 4.789286136627197, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.01885785162448883, + "learning_rate": 0.0006, + "loss": 4.669743537902832, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.018324270844459534, + "learning_rate": 0.0006, + "loss": 4.65901517868042, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.017664821818470955, + "learning_rate": 0.0006, + "loss": 4.667839527130127, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.015702905133366585, + "learning_rate": 0.0006, + "loss": 4.64700984954834, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.016674194484949112, + "learning_rate": 0.0006, + "loss": 4.694093227386475, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.016869477927684784, + "learning_rate": 0.0006, + "loss": 4.619544982910156, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.016557540744543076, + "learning_rate": 0.0006, + "loss": 4.568667411804199, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014500945806503296, + "learning_rate": 0.0006, + "loss": 4.561741828918457, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.014228739775717258, + "learning_rate": 0.0006, + "loss": 4.654277324676514, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.013894238509237766, + "learning_rate": 0.0006, + "loss": 4.5867719650268555, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01550068985670805, + "learning_rate": 0.0006, + "loss": 4.61168098449707, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.015940772369503975, + "learning_rate": 0.0006, + "loss": 4.691247940063477, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.01714535616338253, + "learning_rate": 0.0006, + "loss": 4.61183500289917, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.017925038933753967, + "learning_rate": 0.0006, + "loss": 4.719022750854492, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.018141552805900574, + "learning_rate": 0.0006, + "loss": 4.760281562805176, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.017131328582763672, + "learning_rate": 0.0006, + "loss": 4.709680080413818, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.018730048090219498, + "learning_rate": 0.0006, + "loss": 4.5994720458984375, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.018894298002123833, + "learning_rate": 0.0006, + "loss": 4.669070720672607, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.017643198370933533, + "learning_rate": 0.0006, + "loss": 4.641790390014648, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.017647244036197662, + "learning_rate": 0.0006, + "loss": 4.653899192810059, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.017304198816418648, + "learning_rate": 0.0006, + "loss": 4.7487473487854, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01638447679579258, + "learning_rate": 0.0006, + "loss": 4.71945858001709, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.015892071649432182, + "learning_rate": 0.0006, + "loss": 4.744903564453125, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.017420614138245583, + "learning_rate": 0.0006, + "loss": 4.781563758850098, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.016498520970344543, + "learning_rate": 0.0006, + "loss": 4.622779846191406, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.017482127994298935, + "learning_rate": 0.0006, + "loss": 4.62825870513916, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.019018379971385002, + "learning_rate": 0.0006, + "loss": 4.693194389343262, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01993914321064949, + "learning_rate": 0.0006, + "loss": 4.742903709411621, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02301335707306862, + "learning_rate": 0.0006, + "loss": 4.613251686096191, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.024088222533464432, + "learning_rate": 0.0006, + "loss": 4.594305992126465, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.02493320405483246, + "learning_rate": 0.0006, + "loss": 4.700588226318359, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.02501937560737133, + "learning_rate": 0.0006, + "loss": 4.601742744445801, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.023490751162171364, + "learning_rate": 0.0006, + "loss": 4.589211463928223, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.019695665687322617, + "learning_rate": 0.0006, + "loss": 4.635081768035889, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.019276980310678482, + "learning_rate": 0.0006, + "loss": 4.737002372741699, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.022430511191487312, + "learning_rate": 0.0006, + "loss": 4.808295726776123, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.021543916314840317, + "learning_rate": 0.0006, + "loss": 4.559475421905518, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.019885241985321045, + "learning_rate": 0.0006, + "loss": 4.647948741912842, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.02115175500512123, + "learning_rate": 0.0006, + "loss": 4.680505752563477, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.02156151458621025, + "learning_rate": 0.0006, + "loss": 4.673477649688721, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.02184601128101349, + "learning_rate": 0.0006, + "loss": 4.644402027130127, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.0211983360350132, + "learning_rate": 0.0006, + "loss": 4.821558475494385, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.019636567682027817, + "learning_rate": 0.0006, + "loss": 4.578868865966797, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.019997350871562958, + "learning_rate": 0.0006, + "loss": 4.691307544708252, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.872478008270264, + "eval_runtime": 56.8513, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.018830662593245506, + "learning_rate": 0.0006, + "loss": 4.761779308319092, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.016876481473445892, + "learning_rate": 0.0006, + "loss": 4.61148738861084, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.015550863929092884, + "learning_rate": 0.0006, + "loss": 4.676124095916748, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.017178896814584732, + "learning_rate": 0.0006, + "loss": 4.588935852050781, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.015465447679162025, + "learning_rate": 0.0006, + "loss": 4.605499267578125, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.01606195978820324, + "learning_rate": 0.0006, + "loss": 4.723373889923096, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.016521936282515526, + "learning_rate": 0.0006, + "loss": 4.646608352661133, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.01589205674827099, + "learning_rate": 0.0006, + "loss": 4.498138427734375, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.01524856686592102, + "learning_rate": 0.0006, + "loss": 4.594562530517578, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014714641496539116, + "learning_rate": 0.0006, + "loss": 4.630744934082031, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.015333331190049648, + "learning_rate": 0.0006, + "loss": 4.682388782501221, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.01550073828548193, + "learning_rate": 0.0006, + "loss": 4.692652702331543, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.014783318154513836, + "learning_rate": 0.0006, + "loss": 4.6223835945129395, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.014372751116752625, + "learning_rate": 0.0006, + "loss": 4.52687406539917, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.014779130928218365, + "learning_rate": 0.0006, + "loss": 4.637009620666504, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.015697073191404343, + "learning_rate": 0.0006, + "loss": 4.682981014251709, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.016523664817214012, + "learning_rate": 0.0006, + "loss": 4.705722808837891, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.017633996903896332, + "learning_rate": 0.0006, + "loss": 4.574686050415039, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.019955076277256012, + "learning_rate": 0.0006, + "loss": 4.634450912475586, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.02132618986070156, + "learning_rate": 0.0006, + "loss": 4.56110143661499, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.0207593385130167, + "learning_rate": 0.0006, + "loss": 4.629396438598633, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.017882540822029114, + "learning_rate": 0.0006, + "loss": 4.730963230133057, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01917876861989498, + "learning_rate": 0.0006, + "loss": 4.64585542678833, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01943974196910858, + "learning_rate": 0.0006, + "loss": 4.656181335449219, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.019904715940356255, + "learning_rate": 0.0006, + "loss": 4.619329452514648, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.019174346700310707, + "learning_rate": 0.0006, + "loss": 4.647892475128174, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.018244758248329163, + "learning_rate": 0.0006, + "loss": 4.56166934967041, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.018914205953478813, + "learning_rate": 0.0006, + "loss": 4.648063659667969, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.01961071416735649, + "learning_rate": 0.0006, + "loss": 4.665525436401367, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.01841311901807785, + "learning_rate": 0.0006, + "loss": 4.611342430114746, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.01863427273929119, + "learning_rate": 0.0006, + "loss": 4.664256572723389, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.020506877452135086, + "learning_rate": 0.0006, + "loss": 4.631237506866455, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.02223833277821541, + "learning_rate": 0.0006, + "loss": 4.668654441833496, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.023336702957749367, + "learning_rate": 0.0006, + "loss": 4.562833309173584, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023052413016557693, + "learning_rate": 0.0006, + "loss": 4.621532917022705, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.023420924320816994, + "learning_rate": 0.0006, + "loss": 4.665465354919434, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.024071281775832176, + "learning_rate": 0.0006, + "loss": 4.595980644226074, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.023406682536005974, + "learning_rate": 0.0006, + "loss": 4.642635345458984, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02138647809624672, + "learning_rate": 0.0006, + "loss": 4.704513072967529, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.0221384447067976, + "learning_rate": 0.0006, + "loss": 4.567503452301025, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.020879851654171944, + "learning_rate": 0.0006, + "loss": 4.664057731628418, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.021478967741131783, + "learning_rate": 0.0006, + "loss": 4.539627552032471, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.022748656570911407, + "learning_rate": 0.0006, + "loss": 4.5654497146606445, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.021870248019695282, + "learning_rate": 0.0006, + "loss": 4.56899356842041, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.021711355075240135, + "learning_rate": 0.0006, + "loss": 4.657808303833008, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.022810909897089005, + "learning_rate": 0.0006, + "loss": 4.6526689529418945, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.022561343386769295, + "learning_rate": 0.0006, + "loss": 4.534887790679932, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019909802824258804, + "learning_rate": 0.0006, + "loss": 4.600790023803711, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.020092325285077095, + "learning_rate": 0.0006, + "loss": 4.773830413818359, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.018834445625543594, + "learning_rate": 0.0006, + "loss": 4.628868103027344, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.017552798613905907, + "learning_rate": 0.0006, + "loss": 4.647500038146973, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01751251146197319, + "learning_rate": 0.0006, + "loss": 4.590545654296875, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01699099875986576, + "learning_rate": 0.0006, + "loss": 4.509699821472168, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.015808649361133575, + "learning_rate": 0.0006, + "loss": 4.606992244720459, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01655443385243416, + "learning_rate": 0.0006, + "loss": 4.585219860076904, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01708163693547249, + "learning_rate": 0.0006, + "loss": 4.668147087097168, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.017161913216114044, + "learning_rate": 0.0006, + "loss": 4.7170634269714355, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.017262674868106842, + "learning_rate": 0.0006, + "loss": 4.572132110595703, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.017713576555252075, + "learning_rate": 0.0006, + "loss": 4.702538967132568, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.018730733543634415, + "learning_rate": 0.0006, + "loss": 4.629721641540527, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.017960187047719955, + "learning_rate": 0.0006, + "loss": 4.661039352416992, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.01750079356133938, + "learning_rate": 0.0006, + "loss": 4.503710746765137, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.016630422323942184, + "learning_rate": 0.0006, + "loss": 4.6582136154174805, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.015896832570433617, + "learning_rate": 0.0006, + "loss": 4.596693992614746, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.01566407084465027, + "learning_rate": 0.0006, + "loss": 4.6337199211120605, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.016823330894112587, + "learning_rate": 0.0006, + "loss": 4.686994552612305, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.016208255663514137, + "learning_rate": 0.0006, + "loss": 4.6205549240112305, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.0171707421541214, + "learning_rate": 0.0006, + "loss": 4.582968711853027, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.017009446397423744, + "learning_rate": 0.0006, + "loss": 4.510274887084961, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.01565811224281788, + "learning_rate": 0.0006, + "loss": 4.55834436416626, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.015276037156581879, + "learning_rate": 0.0006, + "loss": 4.537870407104492, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.017305882647633553, + "learning_rate": 0.0006, + "loss": 4.611305236816406, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.775513172149658, + "eval_runtime": 56.7341, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.016752436757087708, + "learning_rate": 0.0006, + "loss": 4.541312217712402, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01643134094774723, + "learning_rate": 0.0006, + "loss": 4.585402965545654, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.01888720504939556, + "learning_rate": 0.0006, + "loss": 4.597825050354004, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.022044632583856583, + "learning_rate": 0.0006, + "loss": 4.532772064208984, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.022986987605690956, + "learning_rate": 0.0006, + "loss": 4.551351547241211, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.020116791129112244, + "learning_rate": 0.0006, + "loss": 4.484149932861328, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.021442247554659843, + "learning_rate": 0.0006, + "loss": 4.515316486358643, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.020974772050976753, + "learning_rate": 0.0006, + "loss": 4.51606559753418, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.021274864673614502, + "learning_rate": 0.0006, + "loss": 4.587329864501953, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.022799784317612648, + "learning_rate": 0.0006, + "loss": 4.568149089813232, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.021784933283925056, + "learning_rate": 0.0006, + "loss": 4.515336990356445, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.024261271581053734, + "learning_rate": 0.0006, + "loss": 4.5127081871032715, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.02248375490307808, + "learning_rate": 0.0006, + "loss": 4.754251003265381, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.02136492356657982, + "learning_rate": 0.0006, + "loss": 4.58193826675415, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.020727066323161125, + "learning_rate": 0.0006, + "loss": 4.564209938049316, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.017789119854569435, + "learning_rate": 0.0006, + "loss": 4.698680877685547, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.017603034153580666, + "learning_rate": 0.0006, + "loss": 4.501171588897705, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.017345862463116646, + "learning_rate": 0.0006, + "loss": 4.593590259552002, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.018047425895929337, + "learning_rate": 0.0006, + "loss": 4.597467422485352, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.01947222836315632, + "learning_rate": 0.0006, + "loss": 4.5072221755981445, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.019598500803112984, + "learning_rate": 0.0006, + "loss": 4.601351737976074, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.020002514123916626, + "learning_rate": 0.0006, + "loss": 4.580431938171387, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.0216530729085207, + "learning_rate": 0.0006, + "loss": 4.424681663513184, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.01949235238134861, + "learning_rate": 0.0006, + "loss": 4.751883029937744, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.017779016867280006, + "learning_rate": 0.0006, + "loss": 4.541490077972412, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.01909121684730053, + "learning_rate": 0.0006, + "loss": 4.594141006469727, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.019894849509000778, + "learning_rate": 0.0006, + "loss": 4.555070400238037, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.01993032731115818, + "learning_rate": 0.0006, + "loss": 4.521401405334473, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.0210788045078516, + "learning_rate": 0.0006, + "loss": 4.672842979431152, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.02274598926305771, + "learning_rate": 0.0006, + "loss": 4.629504203796387, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.020448120310902596, + "learning_rate": 0.0006, + "loss": 4.548209190368652, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.017565643414855003, + "learning_rate": 0.0006, + "loss": 4.529054641723633, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.019504237920045853, + "learning_rate": 0.0006, + "loss": 4.48182487487793, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016823742538690567, + "learning_rate": 0.0006, + "loss": 4.608696937561035, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.01602894812822342, + "learning_rate": 0.0006, + "loss": 4.620062351226807, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.016615642234683037, + "learning_rate": 0.0006, + "loss": 4.69378662109375, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018576730042696, + "learning_rate": 0.0006, + "loss": 4.537041664123535, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.018611151725053787, + "learning_rate": 0.0006, + "loss": 4.496169090270996, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018622413277626038, + "learning_rate": 0.0006, + "loss": 4.557358264923096, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01823544315993786, + "learning_rate": 0.0006, + "loss": 4.730792999267578, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018796846270561218, + "learning_rate": 0.0006, + "loss": 4.781791687011719, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01728292554616928, + "learning_rate": 0.0006, + "loss": 4.515257835388184, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.01699589006602764, + "learning_rate": 0.0006, + "loss": 4.5996904373168945, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.015905970707535744, + "learning_rate": 0.0006, + "loss": 4.601301193237305, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.01338571310043335, + "learning_rate": 0.0006, + "loss": 4.557708740234375, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.015298855490982533, + "learning_rate": 0.0006, + "loss": 4.542287349700928, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.014753367751836777, + "learning_rate": 0.0006, + "loss": 4.595787048339844, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.015527483075857162, + "learning_rate": 0.0006, + "loss": 4.607529163360596, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.015042162500321865, + "learning_rate": 0.0006, + "loss": 4.686626434326172, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.0159393772482872, + "learning_rate": 0.0006, + "loss": 4.47656774520874, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01532787922769785, + "learning_rate": 0.0006, + "loss": 4.55678653717041, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.016332164406776428, + "learning_rate": 0.0006, + "loss": 4.51441764831543, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.018901726230978966, + "learning_rate": 0.0006, + "loss": 4.57227897644043, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.018589433282613754, + "learning_rate": 0.0006, + "loss": 4.550070762634277, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.01634034886956215, + "learning_rate": 0.0006, + "loss": 4.586910724639893, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.017599433660507202, + "learning_rate": 0.0006, + "loss": 4.5891923904418945, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01878446340560913, + "learning_rate": 0.0006, + "loss": 4.57227087020874, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.01803469844162464, + "learning_rate": 0.0006, + "loss": 4.583845138549805, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.018505144864320755, + "learning_rate": 0.0006, + "loss": 4.561151504516602, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.01920081488788128, + "learning_rate": 0.0006, + "loss": 4.506834030151367, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.0177712831646204, + "learning_rate": 0.0006, + "loss": 4.557195663452148, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.019113918766379356, + "learning_rate": 0.0006, + "loss": 4.651227951049805, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.02012539468705654, + "learning_rate": 0.0006, + "loss": 4.4952239990234375, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.019993646070361137, + "learning_rate": 0.0006, + "loss": 4.517433166503906, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.020041441544890404, + "learning_rate": 0.0006, + "loss": 4.489407539367676, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.0183611661195755, + "learning_rate": 0.0006, + "loss": 4.5435638427734375, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.019138429313898087, + "learning_rate": 0.0006, + "loss": 4.545949459075928, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.018554266542196274, + "learning_rate": 0.0006, + "loss": 4.579620361328125, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.0193545650690794, + "learning_rate": 0.0006, + "loss": 4.588255405426025, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.019389165565371513, + "learning_rate": 0.0006, + "loss": 4.520698547363281, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.016601379960775375, + "learning_rate": 0.0006, + "loss": 4.4809064865112305, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.01771303080022335, + "learning_rate": 0.0006, + "loss": 4.721122741699219, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.7903666496276855, + "eval_runtime": 56.9826, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.017720289528369904, + "learning_rate": 0.0006, + "loss": 4.523834705352783, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020137161016464233, + "learning_rate": 0.0006, + "loss": 4.574061870574951, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.02002314105629921, + "learning_rate": 0.0006, + "loss": 4.512407302856445, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.019628183916211128, + "learning_rate": 0.0006, + "loss": 4.502533435821533, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.019808458164334297, + "learning_rate": 0.0006, + "loss": 4.463156700134277, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.021250786259770393, + "learning_rate": 0.0006, + "loss": 4.622682094573975, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.022031694650650024, + "learning_rate": 0.0006, + "loss": 4.451882839202881, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.01898987777531147, + "learning_rate": 0.0006, + "loss": 4.529331684112549, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.018749834969639778, + "learning_rate": 0.0006, + "loss": 4.435815334320068, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01896674931049347, + "learning_rate": 0.0006, + "loss": 4.484029293060303, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018286997452378273, + "learning_rate": 0.0006, + "loss": 4.523510932922363, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.017273059114813805, + "learning_rate": 0.0006, + "loss": 4.451767444610596, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.016499150544404984, + "learning_rate": 0.0006, + "loss": 4.5568647384643555, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.015934469178318977, + "learning_rate": 0.0006, + "loss": 4.4639997482299805, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01714209094643593, + "learning_rate": 0.0006, + "loss": 4.547614097595215, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.019167495891451836, + "learning_rate": 0.0006, + "loss": 4.571097373962402, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.018276996910572052, + "learning_rate": 0.0006, + "loss": 4.489764213562012, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.021296866238117218, + "learning_rate": 0.0006, + "loss": 4.526913642883301, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.022606629878282547, + "learning_rate": 0.0006, + "loss": 4.521624565124512, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.021204711869359016, + "learning_rate": 0.0006, + "loss": 4.500864028930664, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.01988849975168705, + "learning_rate": 0.0006, + "loss": 4.45162296295166, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.020268384367227554, + "learning_rate": 0.0006, + "loss": 4.536198616027832, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.020185677334666252, + "learning_rate": 0.0006, + "loss": 4.440262794494629, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.020106853917241096, + "learning_rate": 0.0006, + "loss": 4.475907325744629, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.018883297219872475, + "learning_rate": 0.0006, + "loss": 4.640661716461182, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.01852620206773281, + "learning_rate": 0.0006, + "loss": 4.578243255615234, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01886451430618763, + "learning_rate": 0.0006, + "loss": 4.546026229858398, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01777452416718006, + "learning_rate": 0.0006, + "loss": 4.521542549133301, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.016626447439193726, + "learning_rate": 0.0006, + "loss": 4.52629280090332, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.01599477231502533, + "learning_rate": 0.0006, + "loss": 4.4796037673950195, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.017021458595991135, + "learning_rate": 0.0006, + "loss": 4.468813419342041, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.016925333067774773, + "learning_rate": 0.0006, + "loss": 4.584898948669434, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.01714998669922352, + "learning_rate": 0.0006, + "loss": 4.523544788360596, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.01802288182079792, + "learning_rate": 0.0006, + "loss": 4.595581531524658, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.0180289875715971, + "learning_rate": 0.0006, + "loss": 4.394440650939941, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.018504245206713676, + "learning_rate": 0.0006, + "loss": 4.4794416427612305, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.019973233342170715, + "learning_rate": 0.0006, + "loss": 4.4241437911987305, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.020063083618879318, + "learning_rate": 0.0006, + "loss": 4.48068380355835, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.02066640742123127, + "learning_rate": 0.0006, + "loss": 4.628993034362793, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.02206416428089142, + "learning_rate": 0.0006, + "loss": 4.564790725708008, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.020640335977077484, + "learning_rate": 0.0006, + "loss": 4.578705787658691, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.019093405455350876, + "learning_rate": 0.0006, + "loss": 4.5023603439331055, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.019459193572402, + "learning_rate": 0.0006, + "loss": 4.594531059265137, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.02021171525120735, + "learning_rate": 0.0006, + "loss": 4.770299911499023, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.021161191165447235, + "learning_rate": 0.0006, + "loss": 4.5380401611328125, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.021013403311371803, + "learning_rate": 0.0006, + "loss": 4.455524921417236, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.020336592569947243, + "learning_rate": 0.0006, + "loss": 4.533672332763672, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.020159434527158737, + "learning_rate": 0.0006, + "loss": 4.596151351928711, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.020432960242033005, + "learning_rate": 0.0006, + "loss": 4.55531644821167, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.025466229766607285, + "learning_rate": 0.0006, + "loss": 4.566912651062012, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.026537558063864708, + "learning_rate": 0.0006, + "loss": 4.526003837585449, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.024397607892751694, + "learning_rate": 0.0006, + "loss": 4.541755676269531, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.022062312811613083, + "learning_rate": 0.0006, + "loss": 4.6532793045043945, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.02204282023012638, + "learning_rate": 0.0006, + "loss": 4.582825660705566, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020844416692852974, + "learning_rate": 0.0006, + "loss": 4.522336483001709, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.019815703853964806, + "learning_rate": 0.0006, + "loss": 4.607398509979248, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016676954925060272, + "learning_rate": 0.0006, + "loss": 4.446206092834473, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.0166452769190073, + "learning_rate": 0.0006, + "loss": 4.505931854248047, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.015928907319903374, + "learning_rate": 0.0006, + "loss": 4.455267429351807, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015219560824334621, + "learning_rate": 0.0006, + "loss": 4.5623674392700195, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.015125355683267117, + "learning_rate": 0.0006, + "loss": 4.554104804992676, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.015612718649208546, + "learning_rate": 0.0006, + "loss": 4.484847068786621, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.015215662308037281, + "learning_rate": 0.0006, + "loss": 4.4802093505859375, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.01511684525758028, + "learning_rate": 0.0006, + "loss": 4.537298679351807, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.013841322623193264, + "learning_rate": 0.0006, + "loss": 4.427488327026367, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.015115964226424694, + "learning_rate": 0.0006, + "loss": 4.4547119140625, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.016411006450653076, + "learning_rate": 0.0006, + "loss": 4.52994966506958, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.017313152551651, + "learning_rate": 0.0006, + "loss": 4.5337653160095215, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.016559796407818794, + "learning_rate": 0.0006, + "loss": 4.5277910232543945, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.015854761004447937, + "learning_rate": 0.0006, + "loss": 4.6334309577941895, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.015658238902688026, + "learning_rate": 0.0006, + "loss": 4.613372802734375, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.0169155802577734, + "learning_rate": 0.0006, + "loss": 4.527356147766113, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.764543056488037, + "eval_runtime": 56.5998, + "eval_samples_per_second": 43.145, + "eval_steps_per_second": 1.36, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.015348963439464569, + "learning_rate": 0.0006, + "loss": 4.466423988342285, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.01768588088452816, + "learning_rate": 0.0006, + "loss": 4.545933246612549, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.017381226643919945, + "learning_rate": 0.0006, + "loss": 4.494060039520264, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.019002335146069527, + "learning_rate": 0.0006, + "loss": 4.395200729370117, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.01827932707965374, + "learning_rate": 0.0006, + "loss": 4.424848556518555, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.016383031383156776, + "learning_rate": 0.0006, + "loss": 4.494722366333008, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016265157610177994, + "learning_rate": 0.0006, + "loss": 4.416522979736328, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.01702386513352394, + "learning_rate": 0.0006, + "loss": 4.600103378295898, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.018495498225092888, + "learning_rate": 0.0006, + "loss": 4.588401794433594, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.017921684309840202, + "learning_rate": 0.0006, + "loss": 4.478075981140137, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.01801278069615364, + "learning_rate": 0.0006, + "loss": 4.413186550140381, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.017211275175213814, + "learning_rate": 0.0006, + "loss": 4.461341857910156, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.017589328810572624, + "learning_rate": 0.0006, + "loss": 4.546969413757324, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.01797698810696602, + "learning_rate": 0.0006, + "loss": 4.530797004699707, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.020408619195222855, + "learning_rate": 0.0006, + "loss": 4.62999153137207, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01994038000702858, + "learning_rate": 0.0006, + "loss": 4.586811065673828, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.019381340593099594, + "learning_rate": 0.0006, + "loss": 4.4769368171691895, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.020484555512666702, + "learning_rate": 0.0006, + "loss": 4.524302959442139, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.02426416426897049, + "learning_rate": 0.0006, + "loss": 4.561705589294434, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.024748101830482483, + "learning_rate": 0.0006, + "loss": 4.4654316902160645, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.020703328773379326, + "learning_rate": 0.0006, + "loss": 4.53033447265625, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.01765458658337593, + "learning_rate": 0.0006, + "loss": 4.5154266357421875, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.018646420910954475, + "learning_rate": 0.0006, + "loss": 4.5224714279174805, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.01907474733889103, + "learning_rate": 0.0006, + "loss": 4.5055437088012695, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.018179846927523613, + "learning_rate": 0.0006, + "loss": 4.462161064147949, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.016697878018021584, + "learning_rate": 0.0006, + "loss": 4.473718643188477, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.017504019662737846, + "learning_rate": 0.0006, + "loss": 4.474348068237305, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.018889913335442543, + "learning_rate": 0.0006, + "loss": 4.4531636238098145, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.019167251884937286, + "learning_rate": 0.0006, + "loss": 4.4278364181518555, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.015895741060376167, + "learning_rate": 0.0006, + "loss": 4.474592208862305, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.01531882956624031, + "learning_rate": 0.0006, + "loss": 4.403757572174072, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.01565849967300892, + "learning_rate": 0.0006, + "loss": 4.507664680480957, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.01602315716445446, + "learning_rate": 0.0006, + "loss": 4.5648908615112305, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.016467366367578506, + "learning_rate": 0.0006, + "loss": 4.428762912750244, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.01604127697646618, + "learning_rate": 0.0006, + "loss": 4.527771472930908, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.018825586885213852, + "learning_rate": 0.0006, + "loss": 4.485386848449707, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.019258007407188416, + "learning_rate": 0.0006, + "loss": 4.414972305297852, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.018192021176218987, + "learning_rate": 0.0006, + "loss": 4.412166595458984, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.016120698302984238, + "learning_rate": 0.0006, + "loss": 4.484317779541016, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.016496622934937477, + "learning_rate": 0.0006, + "loss": 4.514986515045166, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01795116998255253, + "learning_rate": 0.0006, + "loss": 4.41473388671875, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.019062276929616928, + "learning_rate": 0.0006, + "loss": 4.601996898651123, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.017174631357192993, + "learning_rate": 0.0006, + "loss": 4.571218490600586, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.014015606604516506, + "learning_rate": 0.0006, + "loss": 4.526606559753418, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.015358510427176952, + "learning_rate": 0.0006, + "loss": 4.604286193847656, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01619391329586506, + "learning_rate": 0.0006, + "loss": 4.560883522033691, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.015972964465618134, + "learning_rate": 0.0006, + "loss": 4.479755401611328, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016861025243997574, + "learning_rate": 0.0006, + "loss": 4.474471092224121, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01822078227996826, + "learning_rate": 0.0006, + "loss": 4.509489059448242, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.019415007904171944, + "learning_rate": 0.0006, + "loss": 4.508941173553467, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.019856126978993416, + "learning_rate": 0.0006, + "loss": 4.446670055389404, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01842520199716091, + "learning_rate": 0.0006, + "loss": 4.524422645568848, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.017196480184793472, + "learning_rate": 0.0006, + "loss": 4.498466968536377, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.01823010854423046, + "learning_rate": 0.0006, + "loss": 4.534684181213379, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.020398065447807312, + "learning_rate": 0.0006, + "loss": 4.456494331359863, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.022992262616753578, + "learning_rate": 0.0006, + "loss": 4.476374626159668, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.02235420234501362, + "learning_rate": 0.0006, + "loss": 4.587776184082031, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.02093626745045185, + "learning_rate": 0.0006, + "loss": 4.424172401428223, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.01997302658855915, + "learning_rate": 0.0006, + "loss": 4.448066711425781, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.018265459686517715, + "learning_rate": 0.0006, + "loss": 4.436610698699951, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.01673833839595318, + "learning_rate": 0.0006, + "loss": 4.5502214431762695, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.017908833920955658, + "learning_rate": 0.0006, + "loss": 4.550530433654785, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.01796654611825943, + "learning_rate": 0.0006, + "loss": 4.531463146209717, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01646183803677559, + "learning_rate": 0.0006, + "loss": 4.498582363128662, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.015430327504873276, + "learning_rate": 0.0006, + "loss": 4.52571439743042, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.016527952626347542, + "learning_rate": 0.0006, + "loss": 4.547013759613037, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.016263660043478012, + "learning_rate": 0.0006, + "loss": 4.468754291534424, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.016347210854291916, + "learning_rate": 0.0006, + "loss": 4.450741767883301, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.01701398566365242, + "learning_rate": 0.0006, + "loss": 4.534899711608887, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.017693081870675087, + "learning_rate": 0.0006, + "loss": 4.460553169250488, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.018746308982372284, + "learning_rate": 0.0006, + "loss": 4.538792610168457, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020099416375160217, + "learning_rate": 0.0006, + "loss": 4.449617385864258, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.7756571769714355, + "eval_runtime": 57.1804, + "eval_samples_per_second": 42.707, + "eval_steps_per_second": 1.347, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.018726933747529984, + "learning_rate": 0.0006, + "loss": 4.446059703826904, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.01956087350845337, + "learning_rate": 0.0006, + "loss": 4.4520039558410645, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.020092811435461044, + "learning_rate": 0.0006, + "loss": 4.543343544006348, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.01843622885644436, + "learning_rate": 0.0006, + "loss": 4.535579204559326, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.016989488154649734, + "learning_rate": 0.0006, + "loss": 4.52316951751709, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.01805422455072403, + "learning_rate": 0.0006, + "loss": 4.5393266677856445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.01874629408121109, + "learning_rate": 0.0006, + "loss": 4.422059059143066, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01804221048951149, + "learning_rate": 0.0006, + "loss": 4.50508451461792, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.01785474829375744, + "learning_rate": 0.0006, + "loss": 4.412599563598633, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01971791312098503, + "learning_rate": 0.0006, + "loss": 4.518318176269531, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.02061633951961994, + "learning_rate": 0.0006, + "loss": 4.332393169403076, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.020563429221510887, + "learning_rate": 0.0006, + "loss": 4.445863723754883, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.019995713606476784, + "learning_rate": 0.0006, + "loss": 4.497312068939209, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019823070615530014, + "learning_rate": 0.0006, + "loss": 4.475085258483887, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.019039355218410492, + "learning_rate": 0.0006, + "loss": 4.50432014465332, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.017876973375678062, + "learning_rate": 0.0006, + "loss": 4.384788513183594, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.018697699531912804, + "learning_rate": 0.0006, + "loss": 4.410778045654297, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.018826806917786598, + "learning_rate": 0.0006, + "loss": 4.524034023284912, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.01824226602911949, + "learning_rate": 0.0006, + "loss": 4.526393890380859, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.01753208599984646, + "learning_rate": 0.0006, + "loss": 4.493703365325928, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01742432825267315, + "learning_rate": 0.0006, + "loss": 4.392831802368164, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.017099356278777122, + "learning_rate": 0.0006, + "loss": 4.578668594360352, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.017946023494005203, + "learning_rate": 0.0006, + "loss": 4.456718444824219, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.017705943435430527, + "learning_rate": 0.0006, + "loss": 4.553682327270508, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.016647523269057274, + "learning_rate": 0.0006, + "loss": 4.525016784667969, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.015407416969537735, + "learning_rate": 0.0006, + "loss": 4.500231742858887, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.017264019697904587, + "learning_rate": 0.0006, + "loss": 4.4615349769592285, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.01871555484831333, + "learning_rate": 0.0006, + "loss": 4.281026840209961, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.017361650243401527, + "learning_rate": 0.0006, + "loss": 4.479382514953613, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.019781362265348434, + "learning_rate": 0.0006, + "loss": 4.402320861816406, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.019043298438191414, + "learning_rate": 0.0006, + "loss": 4.476983070373535, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017137985676527023, + "learning_rate": 0.0006, + "loss": 4.418689250946045, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.016481833532452583, + "learning_rate": 0.0006, + "loss": 4.579145431518555, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.016507714986801147, + "learning_rate": 0.0006, + "loss": 4.519158363342285, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.018092872574925423, + "learning_rate": 0.0006, + "loss": 4.513748645782471, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.018153440207242966, + "learning_rate": 0.0006, + "loss": 4.645384311676025, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.01657119207084179, + "learning_rate": 0.0006, + "loss": 4.399757385253906, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.0156280305236578, + "learning_rate": 0.0006, + "loss": 4.467836380004883, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.016838403418660164, + "learning_rate": 0.0006, + "loss": 4.529176235198975, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.01743236929178238, + "learning_rate": 0.0006, + "loss": 4.440032482147217, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.01553610060364008, + "learning_rate": 0.0006, + "loss": 4.425588607788086, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.017259271815419197, + "learning_rate": 0.0006, + "loss": 4.405452728271484, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.015568481758236885, + "learning_rate": 0.0006, + "loss": 4.444279193878174, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.015331248752772808, + "learning_rate": 0.0006, + "loss": 4.451846122741699, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.015667378902435303, + "learning_rate": 0.0006, + "loss": 4.533931255340576, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.016177602112293243, + "learning_rate": 0.0006, + "loss": 4.467313289642334, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01733129471540451, + "learning_rate": 0.0006, + "loss": 4.496893882751465, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.016539594158530235, + "learning_rate": 0.0006, + "loss": 4.401961326599121, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.016058551147580147, + "learning_rate": 0.0006, + "loss": 4.503375053405762, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.01715777814388275, + "learning_rate": 0.0006, + "loss": 4.3808369636535645, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.01681119203567505, + "learning_rate": 0.0006, + "loss": 4.5469560623168945, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.01634550653398037, + "learning_rate": 0.0006, + "loss": 4.328197956085205, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.01549347210675478, + "learning_rate": 0.0006, + "loss": 4.534621238708496, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.01654665172100067, + "learning_rate": 0.0006, + "loss": 4.4724225997924805, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.017119025811553, + "learning_rate": 0.0006, + "loss": 4.3897247314453125, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.01603223942220211, + "learning_rate": 0.0006, + "loss": 4.412731170654297, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.016774067655205727, + "learning_rate": 0.0006, + "loss": 4.54707145690918, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.016453025862574577, + "learning_rate": 0.0006, + "loss": 4.462501049041748, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.01597629487514496, + "learning_rate": 0.0006, + "loss": 4.229758262634277, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.017999034374952316, + "learning_rate": 0.0006, + "loss": 4.389263153076172, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.02165898121893406, + "learning_rate": 0.0006, + "loss": 4.342130661010742, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.02222822792828083, + "learning_rate": 0.0006, + "loss": 4.492661476135254, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.019378066062927246, + "learning_rate": 0.0006, + "loss": 4.4614105224609375, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.019908083602786064, + "learning_rate": 0.0006, + "loss": 4.370929718017578, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.022296471521258354, + "learning_rate": 0.0006, + "loss": 4.438320159912109, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.021810343489050865, + "learning_rate": 0.0006, + "loss": 4.462067127227783, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.02129209227859974, + "learning_rate": 0.0006, + "loss": 4.541261196136475, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.02079629711806774, + "learning_rate": 0.0006, + "loss": 4.36712646484375, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.022501740604639053, + "learning_rate": 0.0006, + "loss": 4.5279083251953125, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.019121676683425903, + "learning_rate": 0.0006, + "loss": 4.477789878845215, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017429588362574577, + "learning_rate": 0.0006, + "loss": 4.4023847579956055, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.018857665359973907, + "learning_rate": 0.0006, + "loss": 4.494743347167969, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.694555759429932, + "eval_runtime": 56.4146, + "eval_samples_per_second": 43.287, + "eval_steps_per_second": 1.365, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.016505861654877663, + "learning_rate": 0.0006, + "loss": 4.506229877471924, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.017571842297911644, + "learning_rate": 0.0006, + "loss": 4.384073257446289, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.019398365169763565, + "learning_rate": 0.0006, + "loss": 4.482177734375, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.02011062018573284, + "learning_rate": 0.0006, + "loss": 4.440773963928223, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.019026953727006912, + "learning_rate": 0.0006, + "loss": 4.507168769836426, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.020506031811237335, + "learning_rate": 0.0006, + "loss": 4.469079971313477, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.019608162343502045, + "learning_rate": 0.0006, + "loss": 4.395927429199219, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.019419124349951744, + "learning_rate": 0.0006, + "loss": 4.484235763549805, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.021775128319859505, + "learning_rate": 0.0006, + "loss": 4.333173751831055, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.025512272492051125, + "learning_rate": 0.0006, + "loss": 4.460362434387207, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.025157401338219643, + "learning_rate": 0.0006, + "loss": 4.4361395835876465, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.019862616434693336, + "learning_rate": 0.0006, + "loss": 4.426807403564453, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.01932152360677719, + "learning_rate": 0.0006, + "loss": 4.375677108764648, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.01760999858379364, + "learning_rate": 0.0006, + "loss": 4.42259407043457, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.016998112201690674, + "learning_rate": 0.0006, + "loss": 4.481280326843262, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017199190333485603, + "learning_rate": 0.0006, + "loss": 4.496712684631348, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.016833679750561714, + "learning_rate": 0.0006, + "loss": 4.40736722946167, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.017546730116009712, + "learning_rate": 0.0006, + "loss": 4.4915266036987305, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.016595078632235527, + "learning_rate": 0.0006, + "loss": 4.460815906524658, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.017728568986058235, + "learning_rate": 0.0006, + "loss": 4.346048831939697, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.01849197782576084, + "learning_rate": 0.0006, + "loss": 4.399681091308594, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.018336588516831398, + "learning_rate": 0.0006, + "loss": 4.414183616638184, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.018086519092321396, + "learning_rate": 0.0006, + "loss": 4.346586227416992, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.019697299227118492, + "learning_rate": 0.0006, + "loss": 4.4420952796936035, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.02198074571788311, + "learning_rate": 0.0006, + "loss": 4.588729381561279, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.022813329473137856, + "learning_rate": 0.0006, + "loss": 4.49729585647583, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.022006649523973465, + "learning_rate": 0.0006, + "loss": 4.475641250610352, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.023188438266515732, + "learning_rate": 0.0006, + "loss": 4.404322624206543, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.022226877510547638, + "learning_rate": 0.0006, + "loss": 4.50904655456543, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.021082637831568718, + "learning_rate": 0.0006, + "loss": 4.339011192321777, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.02013542130589485, + "learning_rate": 0.0006, + "loss": 4.516942024230957, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.01898968778550625, + "learning_rate": 0.0006, + "loss": 4.315804481506348, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.019450880587100983, + "learning_rate": 0.0006, + "loss": 4.4794921875, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.018577802926301956, + "learning_rate": 0.0006, + "loss": 4.422323703765869, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.01778412237763405, + "learning_rate": 0.0006, + "loss": 4.429789066314697, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.017134476453065872, + "learning_rate": 0.0006, + "loss": 4.416459083557129, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.017153557389974594, + "learning_rate": 0.0006, + "loss": 4.446170806884766, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.014998532831668854, + "learning_rate": 0.0006, + "loss": 4.419447898864746, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.015877850353717804, + "learning_rate": 0.0006, + "loss": 4.312338829040527, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.01569489948451519, + "learning_rate": 0.0006, + "loss": 4.572905540466309, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01591474376618862, + "learning_rate": 0.0006, + "loss": 4.414907455444336, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.015395903028547764, + "learning_rate": 0.0006, + "loss": 4.549131393432617, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.015715358778834343, + "learning_rate": 0.0006, + "loss": 4.530194282531738, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.016042975708842278, + "learning_rate": 0.0006, + "loss": 4.322149276733398, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.01664186641573906, + "learning_rate": 0.0006, + "loss": 4.424866676330566, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.01631534844636917, + "learning_rate": 0.0006, + "loss": 4.441752910614014, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.014057058840990067, + "learning_rate": 0.0006, + "loss": 4.399982452392578, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.015076273120939732, + "learning_rate": 0.0006, + "loss": 4.441324234008789, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.014138277620077133, + "learning_rate": 0.0006, + "loss": 4.4312543869018555, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.013762440532445908, + "learning_rate": 0.0006, + "loss": 4.387807846069336, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.015058411285281181, + "learning_rate": 0.0006, + "loss": 4.594747543334961, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.434604167938232, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.015315295197069645, + "learning_rate": 0.0006, + "loss": 4.588066101074219, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.018678732216358185, + "learning_rate": 0.0006, + "loss": 4.3514204025268555, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.019521350041031837, + "learning_rate": 0.0006, + "loss": 4.437658786773682, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.018363403156399727, + "learning_rate": 0.0006, + "loss": 4.495002269744873, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.016830725595355034, + "learning_rate": 0.0006, + "loss": 4.402471542358398, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.01722072623670101, + "learning_rate": 0.0006, + "loss": 4.437653541564941, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.017459599301218987, + "learning_rate": 0.0006, + "loss": 4.366603851318359, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.018902219831943512, + "learning_rate": 0.0006, + "loss": 4.501561164855957, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.020814096555113792, + "learning_rate": 0.0006, + "loss": 4.535295486450195, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.018990976735949516, + "learning_rate": 0.0006, + "loss": 4.333580017089844, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.01837880350649357, + "learning_rate": 0.0006, + "loss": 4.313388824462891, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.019820360466837883, + "learning_rate": 0.0006, + "loss": 4.388157844543457, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.020181143656373024, + "learning_rate": 0.0006, + "loss": 4.572224140167236, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.021206077188253403, + "learning_rate": 0.0006, + "loss": 4.3127970695495605, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.017962060868740082, + "learning_rate": 0.0006, + "loss": 4.384734630584717, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.01737220585346222, + "learning_rate": 0.0006, + "loss": 4.4075751304626465, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.01839268207550049, + "learning_rate": 0.0006, + "loss": 4.553779125213623, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.020102640613913536, + "learning_rate": 0.0006, + "loss": 4.399398326873779, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.018831463530659676, + "learning_rate": 0.0006, + "loss": 4.358499526977539, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.018254734575748444, + "learning_rate": 0.0006, + "loss": 4.384119510650635, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.7524237632751465, + "eval_runtime": 57.8812, + "eval_samples_per_second": 42.19, + "eval_steps_per_second": 1.33, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.019728587940335274, + "learning_rate": 0.0006, + "loss": 4.32363224029541, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.02216625027358532, + "learning_rate": 0.0006, + "loss": 4.436988353729248, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.02227606810629368, + "learning_rate": 0.0006, + "loss": 4.406063556671143, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.020187009125947952, + "learning_rate": 0.0006, + "loss": 4.324402332305908, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.02095157466828823, + "learning_rate": 0.0006, + "loss": 4.456544876098633, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.02146013453602791, + "learning_rate": 0.0006, + "loss": 4.368387222290039, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.02162035182118416, + "learning_rate": 0.0006, + "loss": 4.48838996887207, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.021745676174759865, + "learning_rate": 0.0006, + "loss": 4.30778694152832, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.02173588052392006, + "learning_rate": 0.0006, + "loss": 4.433149814605713, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.020950648933649063, + "learning_rate": 0.0006, + "loss": 4.520089149475098, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.020684752613306046, + "learning_rate": 0.0006, + "loss": 4.372437477111816, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.019472869113087654, + "learning_rate": 0.0006, + "loss": 4.437015533447266, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.01846962980926037, + "learning_rate": 0.0006, + "loss": 4.449531555175781, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.017377803102135658, + "learning_rate": 0.0006, + "loss": 4.458279132843018, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.016002200543880463, + "learning_rate": 0.0006, + "loss": 4.266297817230225, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.015575903467833996, + "learning_rate": 0.0006, + "loss": 4.342632293701172, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.015634698793292046, + "learning_rate": 0.0006, + "loss": 4.338623523712158, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.01449581328779459, + "learning_rate": 0.0006, + "loss": 4.354330062866211, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.01487616915255785, + "learning_rate": 0.0006, + "loss": 4.486835479736328, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.016346555203199387, + "learning_rate": 0.0006, + "loss": 4.407098770141602, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.01682182401418686, + "learning_rate": 0.0006, + "loss": 4.436809539794922, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.017704004421830177, + "learning_rate": 0.0006, + "loss": 4.3854780197143555, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.018437420949339867, + "learning_rate": 0.0006, + "loss": 4.413283824920654, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.016911081969738007, + "learning_rate": 0.0006, + "loss": 4.357306003570557, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.01642223633825779, + "learning_rate": 0.0006, + "loss": 4.422322750091553, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.01780877821147442, + "learning_rate": 0.0006, + "loss": 4.387620449066162, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.01653975248336792, + "learning_rate": 0.0006, + "loss": 4.327267646789551, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.01572156324982643, + "learning_rate": 0.0006, + "loss": 4.475979804992676, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.017156066372990608, + "learning_rate": 0.0006, + "loss": 4.4011125564575195, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.01788829080760479, + "learning_rate": 0.0006, + "loss": 4.4174089431762695, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.01728985086083412, + "learning_rate": 0.0006, + "loss": 4.429244041442871, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.016534466296434402, + "learning_rate": 0.0006, + "loss": 4.326068878173828, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.016582584008574486, + "learning_rate": 0.0006, + "loss": 4.2975969314575195, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.018034178763628006, + "learning_rate": 0.0006, + "loss": 4.501319408416748, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.017827702686190605, + "learning_rate": 0.0006, + "loss": 4.502284526824951, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.017182299867272377, + "learning_rate": 0.0006, + "loss": 4.380828857421875, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.019101588055491447, + "learning_rate": 0.0006, + "loss": 4.31564998626709, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.0204677302390337, + "learning_rate": 0.0006, + "loss": 4.3840131759643555, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.0188993439078331, + "learning_rate": 0.0006, + "loss": 4.4257731437683105, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.01730230636894703, + "learning_rate": 0.0006, + "loss": 4.445132255554199, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.01801574043929577, + "learning_rate": 0.0006, + "loss": 4.427064895629883, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.018417565152049065, + "learning_rate": 0.0006, + "loss": 4.319595813751221, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.01683160290122032, + "learning_rate": 0.0006, + "loss": 4.349218368530273, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.016973106190562248, + "learning_rate": 0.0006, + "loss": 4.375990390777588, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.015688583254814148, + "learning_rate": 0.0006, + "loss": 4.408170223236084, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.01573043316602707, + "learning_rate": 0.0006, + "loss": 4.416255950927734, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.01744293048977852, + "learning_rate": 0.0006, + "loss": 4.477141380310059, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.017503513023257256, + "learning_rate": 0.0006, + "loss": 4.3367486000061035, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.016592150554060936, + "learning_rate": 0.0006, + "loss": 4.314785480499268, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.015774084255099297, + "learning_rate": 0.0006, + "loss": 4.448209285736084, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.015779150649905205, + "learning_rate": 0.0006, + "loss": 4.392690658569336, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.015392648056149483, + "learning_rate": 0.0006, + "loss": 4.437741756439209, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.01530836895108223, + "learning_rate": 0.0006, + "loss": 4.424595832824707, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.015600779093801975, + "learning_rate": 0.0006, + "loss": 4.335417747497559, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.014646518044173717, + "learning_rate": 0.0006, + "loss": 4.279081344604492, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.014132864773273468, + "learning_rate": 0.0006, + "loss": 4.345950126647949, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.01425724383443594, + "learning_rate": 0.0006, + "loss": 4.277588844299316, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.015396572649478912, + "learning_rate": 0.0006, + "loss": 4.326196670532227, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.01683669537305832, + "learning_rate": 0.0006, + "loss": 4.3294172286987305, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.016365647315979004, + "learning_rate": 0.0006, + "loss": 4.523321628570557, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.017142774537205696, + "learning_rate": 0.0006, + "loss": 4.475842475891113, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.015950839966535568, + "learning_rate": 0.0006, + "loss": 4.397187232971191, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.016180362552404404, + "learning_rate": 0.0006, + "loss": 4.414778709411621, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.017615556716918945, + "learning_rate": 0.0006, + "loss": 4.348567962646484, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.01896277628839016, + "learning_rate": 0.0006, + "loss": 4.431762218475342, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.01990971714258194, + "learning_rate": 0.0006, + "loss": 4.431285858154297, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.0201258547604084, + "learning_rate": 0.0006, + "loss": 4.309872627258301, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.019620198756456375, + "learning_rate": 0.0006, + "loss": 4.484044551849365, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.01771123707294464, + "learning_rate": 0.0006, + "loss": 4.368093013763428, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.017501654103398323, + "learning_rate": 0.0006, + "loss": 4.372651100158691, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.01693608984351158, + "learning_rate": 0.0006, + "loss": 4.379167556762695, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.02002849616110325, + "learning_rate": 0.0006, + "loss": 4.329627990722656, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.723696231842041, + "eval_runtime": 56.1758, + "eval_samples_per_second": 43.471, + "eval_steps_per_second": 1.371, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.02030305378139019, + "learning_rate": 0.0006, + "loss": 4.355566024780273, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.018638934940099716, + "learning_rate": 0.0006, + "loss": 4.314009189605713, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.019917072728276253, + "learning_rate": 0.0006, + "loss": 4.367525577545166, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.020785167813301086, + "learning_rate": 0.0006, + "loss": 4.375561714172363, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.021214453503489494, + "learning_rate": 0.0006, + "loss": 4.418305397033691, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.019051678478717804, + "learning_rate": 0.0006, + "loss": 4.215126991271973, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.02323366142809391, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.026186738163232803, + "learning_rate": 0.0006, + "loss": 4.408508777618408, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.027361227199435234, + "learning_rate": 0.0006, + "loss": 4.329322814941406, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.026564646512269974, + "learning_rate": 0.0006, + "loss": 4.346642017364502, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.025073667988181114, + "learning_rate": 0.0006, + "loss": 4.3984198570251465, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.02347794733941555, + "learning_rate": 0.0006, + "loss": 4.352538108825684, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.02203468047082424, + "learning_rate": 0.0006, + "loss": 4.298148155212402, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.020621957257390022, + "learning_rate": 0.0006, + "loss": 4.480309963226318, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.023710055276751518, + "learning_rate": 0.0006, + "loss": 4.339317321777344, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.0233482476323843, + "learning_rate": 0.0006, + "loss": 4.436896324157715, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.020122403278946877, + "learning_rate": 0.0006, + "loss": 4.236002445220947, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.01957266591489315, + "learning_rate": 0.0006, + "loss": 4.497328281402588, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.018964895978569984, + "learning_rate": 0.0006, + "loss": 4.36360502243042, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.018902769312262535, + "learning_rate": 0.0006, + "loss": 4.427093505859375, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.018918119370937347, + "learning_rate": 0.0006, + "loss": 4.294950485229492, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.019315775483846664, + "learning_rate": 0.0006, + "loss": 4.427952289581299, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.018350522965192795, + "learning_rate": 0.0006, + "loss": 4.272945404052734, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01759534701704979, + "learning_rate": 0.0006, + "loss": 4.348023891448975, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.01646057888865471, + "learning_rate": 0.0006, + "loss": 4.278518199920654, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.015875060111284256, + "learning_rate": 0.0006, + "loss": 4.423527240753174, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.01638529635965824, + "learning_rate": 0.0006, + "loss": 4.321723937988281, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.016942061483860016, + "learning_rate": 0.0006, + "loss": 4.346164703369141, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.015209621749818325, + "learning_rate": 0.0006, + "loss": 4.434340000152588, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.014423009008169174, + "learning_rate": 0.0006, + "loss": 4.342560768127441, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.014369525946676731, + "learning_rate": 0.0006, + "loss": 4.296586990356445, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.01392540242522955, + "learning_rate": 0.0006, + "loss": 4.433176040649414, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.013206899166107178, + "learning_rate": 0.0006, + "loss": 4.32962703704834, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.013427077792584896, + "learning_rate": 0.0006, + "loss": 4.420437335968018, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.014194196090102196, + "learning_rate": 0.0006, + "loss": 4.410478591918945, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.01362364087253809, + "learning_rate": 0.0006, + "loss": 4.338932037353516, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.014299589209258556, + "learning_rate": 0.0006, + "loss": 4.34089994430542, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.014061295427381992, + "learning_rate": 0.0006, + "loss": 4.29201602935791, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.013945229351520538, + "learning_rate": 0.0006, + "loss": 4.326052188873291, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.01387068536132574, + "learning_rate": 0.0006, + "loss": 4.393521308898926, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.013931898400187492, + "learning_rate": 0.0006, + "loss": 4.3739752769470215, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.01497740764170885, + "learning_rate": 0.0006, + "loss": 4.3991618156433105, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.015842905268073082, + "learning_rate": 0.0006, + "loss": 4.405340671539307, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.01580633595585823, + "learning_rate": 0.0006, + "loss": 4.280102729797363, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.016567382961511612, + "learning_rate": 0.0006, + "loss": 4.407657146453857, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01648622378706932, + "learning_rate": 0.0006, + "loss": 4.392590522766113, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.017066307365894318, + "learning_rate": 0.0006, + "loss": 4.324862003326416, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01637062057852745, + "learning_rate": 0.0006, + "loss": 4.34598445892334, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.018038960173726082, + "learning_rate": 0.0006, + "loss": 4.320711612701416, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.018096428364515305, + "learning_rate": 0.0006, + "loss": 4.536527156829834, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.017889291048049927, + "learning_rate": 0.0006, + "loss": 4.388131141662598, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.018771521747112274, + "learning_rate": 0.0006, + "loss": 4.357668876647949, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.02100873365998268, + "learning_rate": 0.0006, + "loss": 4.293972015380859, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.017715785652399063, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.015546333976089954, + "learning_rate": 0.0006, + "loss": 4.311728000640869, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.01605990156531334, + "learning_rate": 0.0006, + "loss": 4.499123573303223, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.016642142087221146, + "learning_rate": 0.0006, + "loss": 4.4319257736206055, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.016120387241244316, + "learning_rate": 0.0006, + "loss": 4.4373650550842285, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.015174890868365765, + "learning_rate": 0.0006, + "loss": 4.314082145690918, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.01551714539527893, + "learning_rate": 0.0006, + "loss": 4.360363960266113, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.014966742135584354, + "learning_rate": 0.0006, + "loss": 4.346253395080566, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.014288028702139854, + "learning_rate": 0.0006, + "loss": 4.464366436004639, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.015291682444512844, + "learning_rate": 0.0006, + "loss": 4.504101753234863, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.014534426853060722, + "learning_rate": 0.0006, + "loss": 4.462172031402588, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.014129819348454475, + "learning_rate": 0.0006, + "loss": 4.350126266479492, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.014932668767869473, + "learning_rate": 0.0006, + "loss": 4.374392032623291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.014647853560745716, + "learning_rate": 0.0006, + "loss": 4.252264499664307, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.013553360477089882, + "learning_rate": 0.0006, + "loss": 4.379729270935059, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.014606194570660591, + "learning_rate": 0.0006, + "loss": 4.363372802734375, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.015000631101429462, + "learning_rate": 0.0006, + "loss": 4.408048629760742, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.015370816923677921, + "learning_rate": 0.0006, + "loss": 4.297451019287109, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.01709144562482834, + "learning_rate": 0.0006, + "loss": 4.294776916503906, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.678618907928467, + "eval_runtime": 56.1668, + "eval_samples_per_second": 43.478, + "eval_steps_per_second": 1.371, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.01727290451526642, + "learning_rate": 0.0006, + "loss": 4.311701774597168, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.020645759999752045, + "learning_rate": 0.0006, + "loss": 4.429897308349609, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.02204226516187191, + "learning_rate": 0.0006, + "loss": 4.365151882171631, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.023069269955158234, + "learning_rate": 0.0006, + "loss": 4.3710832595825195, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.024151179939508438, + "learning_rate": 0.0006, + "loss": 4.403108596801758, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.02511233650147915, + "learning_rate": 0.0006, + "loss": 4.414128303527832, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.025802113115787506, + "learning_rate": 0.0006, + "loss": 4.388265609741211, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.02377348765730858, + "learning_rate": 0.0006, + "loss": 4.357505798339844, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.020876651629805565, + "learning_rate": 0.0006, + "loss": 4.23043966293335, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.01959727331995964, + "learning_rate": 0.0006, + "loss": 4.280970096588135, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.018678616732358932, + "learning_rate": 0.0006, + "loss": 4.346716403961182, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.02108944021165371, + "learning_rate": 0.0006, + "loss": 4.277904510498047, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.020221566781401634, + "learning_rate": 0.0006, + "loss": 4.367693901062012, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.019599711522459984, + "learning_rate": 0.0006, + "loss": 4.257092475891113, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.019833361729979515, + "learning_rate": 0.0006, + "loss": 4.540737628936768, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.020352095365524292, + "learning_rate": 0.0006, + "loss": 4.360923767089844, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.01939748041331768, + "learning_rate": 0.0006, + "loss": 4.393110275268555, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.01884354278445244, + "learning_rate": 0.0006, + "loss": 4.383938789367676, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.02008470520377159, + "learning_rate": 0.0006, + "loss": 4.346344947814941, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.019065195694565773, + "learning_rate": 0.0006, + "loss": 4.36611795425415, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.01938541606068611, + "learning_rate": 0.0006, + "loss": 4.4009599685668945, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.019044723361730576, + "learning_rate": 0.0006, + "loss": 4.340169906616211, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.018850378692150116, + "learning_rate": 0.0006, + "loss": 4.37674617767334, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.019076095893979073, + "learning_rate": 0.0006, + "loss": 4.226874351501465, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.017543038353323936, + "learning_rate": 0.0006, + "loss": 4.337751388549805, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.017170218750834465, + "learning_rate": 0.0006, + "loss": 4.282512664794922, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.01574796624481678, + "learning_rate": 0.0006, + "loss": 4.30638313293457, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.015160846523940563, + "learning_rate": 0.0006, + "loss": 4.336716175079346, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01534526702016592, + "learning_rate": 0.0006, + "loss": 4.2174787521362305, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.015289999544620514, + "learning_rate": 0.0006, + "loss": 4.272546291351318, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.014894185587763786, + "learning_rate": 0.0006, + "loss": 4.296045303344727, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.014869502745568752, + "learning_rate": 0.0006, + "loss": 4.425461769104004, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.01474483497440815, + "learning_rate": 0.0006, + "loss": 4.361867427825928, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.015299245715141296, + "learning_rate": 0.0006, + "loss": 4.364523410797119, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.015201340429484844, + "learning_rate": 0.0006, + "loss": 4.285182952880859, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.016057293862104416, + "learning_rate": 0.0006, + "loss": 4.284873962402344, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.015871062874794006, + "learning_rate": 0.0006, + "loss": 4.326470851898193, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.01586288772523403, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.016092827543616295, + "learning_rate": 0.0006, + "loss": 4.3258867263793945, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.01582016795873642, + "learning_rate": 0.0006, + "loss": 4.358892440795898, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.01688452623784542, + "learning_rate": 0.0006, + "loss": 4.280842304229736, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.015865886583924294, + "learning_rate": 0.0006, + "loss": 4.245567321777344, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.015303662046790123, + "learning_rate": 0.0006, + "loss": 4.383112907409668, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.015649516135454178, + "learning_rate": 0.0006, + "loss": 4.519082069396973, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.0167120099067688, + "learning_rate": 0.0006, + "loss": 4.447847366333008, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.016368716955184937, + "learning_rate": 0.0006, + "loss": 4.434186935424805, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.017379503697156906, + "learning_rate": 0.0006, + "loss": 4.317594528198242, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.018573811277747154, + "learning_rate": 0.0006, + "loss": 4.205793857574463, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.01749090477824211, + "learning_rate": 0.0006, + "loss": 4.3985700607299805, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.0165668074041605, + "learning_rate": 0.0006, + "loss": 4.329226970672607, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.016913846135139465, + "learning_rate": 0.0006, + "loss": 4.388485908508301, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.016813859343528748, + "learning_rate": 0.0006, + "loss": 4.3395915031433105, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.01676975190639496, + "learning_rate": 0.0006, + "loss": 4.254745960235596, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.016370350494980812, + "learning_rate": 0.0006, + "loss": 4.2612223625183105, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.01696198247373104, + "learning_rate": 0.0006, + "loss": 4.327524662017822, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.0189108457416296, + "learning_rate": 0.0006, + "loss": 4.325839519500732, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.01862351782619953, + "learning_rate": 0.0006, + "loss": 4.333662509918213, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.01655014418065548, + "learning_rate": 0.0006, + "loss": 4.39305305480957, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.015899376943707466, + "learning_rate": 0.0006, + "loss": 4.390053749084473, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.016519056633114815, + "learning_rate": 0.0006, + "loss": 4.255903244018555, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.0167539119720459, + "learning_rate": 0.0006, + "loss": 4.255727767944336, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.0171353816986084, + "learning_rate": 0.0006, + "loss": 4.376740455627441, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.016719117760658264, + "learning_rate": 0.0006, + "loss": 4.3376688957214355, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.01561494916677475, + "learning_rate": 0.0006, + "loss": 4.3345746994018555, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.016303174197673798, + "learning_rate": 0.0006, + "loss": 4.332371234893799, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.016722865402698517, + "learning_rate": 0.0006, + "loss": 4.267125129699707, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.017072124406695366, + "learning_rate": 0.0006, + "loss": 4.322751998901367, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.01699932850897312, + "learning_rate": 0.0006, + "loss": 4.378422737121582, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.01742720976471901, + "learning_rate": 0.0006, + "loss": 4.345083236694336, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.018797501921653748, + "learning_rate": 0.0006, + "loss": 4.290443420410156, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.018090683966875076, + "learning_rate": 0.0006, + "loss": 4.377580642700195, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.017305459827184677, + "learning_rate": 0.0006, + "loss": 4.364426612854004, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.647762298583984, + "eval_runtime": 56.5402, + "eval_samples_per_second": 43.191, + "eval_steps_per_second": 1.362, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.016493607312440872, + "learning_rate": 0.0006, + "loss": 4.252255916595459, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.016529450193047523, + "learning_rate": 0.0006, + "loss": 4.352746963500977, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.016940191388130188, + "learning_rate": 0.0006, + "loss": 4.301870822906494, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.01752558909356594, + "learning_rate": 0.0006, + "loss": 4.2762451171875, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.017112884670495987, + "learning_rate": 0.0006, + "loss": 4.2659149169921875, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.01900586113333702, + "learning_rate": 0.0006, + "loss": 4.306058883666992, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.019447680562734604, + "learning_rate": 0.0006, + "loss": 4.450128555297852, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.01948186755180359, + "learning_rate": 0.0006, + "loss": 4.202869415283203, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.01716383546590805, + "learning_rate": 0.0006, + "loss": 4.318876266479492, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.017049988731741905, + "learning_rate": 0.0006, + "loss": 4.237238883972168, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.015169923193752766, + "learning_rate": 0.0006, + "loss": 4.289426803588867, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.014760667458176613, + "learning_rate": 0.0006, + "loss": 4.356924057006836, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.015127750113606453, + "learning_rate": 0.0006, + "loss": 4.142916679382324, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.01549555640667677, + "learning_rate": 0.0006, + "loss": 4.345563888549805, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.016393091529607773, + "learning_rate": 0.0006, + "loss": 4.417495250701904, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.01785266026854515, + "learning_rate": 0.0006, + "loss": 4.3553547859191895, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.017574617639183998, + "learning_rate": 0.0006, + "loss": 4.241024494171143, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.019060306251049042, + "learning_rate": 0.0006, + "loss": 4.3685150146484375, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.0214633010327816, + "learning_rate": 0.0006, + "loss": 4.323866844177246, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.021055003628134727, + "learning_rate": 0.0006, + "loss": 4.278757095336914, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.018618572503328323, + "learning_rate": 0.0006, + "loss": 4.28849983215332, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.018666435033082962, + "learning_rate": 0.0006, + "loss": 4.368590354919434, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.020626146346330643, + "learning_rate": 0.0006, + "loss": 4.2951836585998535, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.01929759792983532, + "learning_rate": 0.0006, + "loss": 4.3664703369140625, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.020956002175807953, + "learning_rate": 0.0006, + "loss": 4.345308303833008, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.019227957352995872, + "learning_rate": 0.0006, + "loss": 4.349849224090576, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.015970097854733467, + "learning_rate": 0.0006, + "loss": 4.223012447357178, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.015951646491885185, + "learning_rate": 0.0006, + "loss": 4.236420631408691, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.016032515093684196, + "learning_rate": 0.0006, + "loss": 4.226899147033691, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01573132909834385, + "learning_rate": 0.0006, + "loss": 4.315916061401367, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.015367005951702595, + "learning_rate": 0.0006, + "loss": 4.358939170837402, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.01669907011091709, + "learning_rate": 0.0006, + "loss": 4.232147216796875, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.018708152696490288, + "learning_rate": 0.0006, + "loss": 4.329852104187012, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.019035987555980682, + "learning_rate": 0.0006, + "loss": 4.260340213775635, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.01832144521176815, + "learning_rate": 0.0006, + "loss": 4.359543800354004, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.015313294716179371, + "learning_rate": 0.0006, + "loss": 4.279088020324707, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.015803923830389977, + "learning_rate": 0.0006, + "loss": 4.230682849884033, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.016748374328017235, + "learning_rate": 0.0006, + "loss": 4.349661827087402, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.015732428058981895, + "learning_rate": 0.0006, + "loss": 4.372441291809082, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.016564829275012016, + "learning_rate": 0.0006, + "loss": 4.323090553283691, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.01603010483086109, + "learning_rate": 0.0006, + "loss": 4.347898006439209, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.01674809865653515, + "learning_rate": 0.0006, + "loss": 4.269165515899658, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.017333930358290672, + "learning_rate": 0.0006, + "loss": 4.353298187255859, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.017519580200314522, + "learning_rate": 0.0006, + "loss": 4.265199661254883, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.017675306648015976, + "learning_rate": 0.0006, + "loss": 4.322596549987793, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.017166534438729286, + "learning_rate": 0.0006, + "loss": 4.413900375366211, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.016362346708774567, + "learning_rate": 0.0006, + "loss": 4.307236671447754, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.016086872667074203, + "learning_rate": 0.0006, + "loss": 4.235316276550293, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.01689537614583969, + "learning_rate": 0.0006, + "loss": 4.390591621398926, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.016836725175380707, + "learning_rate": 0.0006, + "loss": 4.2990031242370605, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.015627335757017136, + "learning_rate": 0.0006, + "loss": 4.166745185852051, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.015007016249001026, + "learning_rate": 0.0006, + "loss": 4.295273303985596, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.016084209084510803, + "learning_rate": 0.0006, + "loss": 4.342424392700195, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.01595073565840721, + "learning_rate": 0.0006, + "loss": 4.3060102462768555, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.016113461926579475, + "learning_rate": 0.0006, + "loss": 4.267908096313477, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.01564556173980236, + "learning_rate": 0.0006, + "loss": 4.324642658233643, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.01575268618762493, + "learning_rate": 0.0006, + "loss": 4.328181743621826, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.016830824315547943, + "learning_rate": 0.0006, + "loss": 4.247906684875488, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.017989547923207283, + "learning_rate": 0.0006, + "loss": 4.228281021118164, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.016783086583018303, + "learning_rate": 0.0006, + "loss": 4.2880964279174805, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.016467561945319176, + "learning_rate": 0.0006, + "loss": 4.36118745803833, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.01780523732304573, + "learning_rate": 0.0006, + "loss": 4.172321319580078, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.017102031037211418, + "learning_rate": 0.0006, + "loss": 4.352826118469238, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.016409218311309814, + "learning_rate": 0.0006, + "loss": 4.212162017822266, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.018660584464669228, + "learning_rate": 0.0006, + "loss": 4.346684455871582, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.017389440909028053, + "learning_rate": 0.0006, + "loss": 4.197913646697998, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.01596485823392868, + "learning_rate": 0.0006, + "loss": 4.370617866516113, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.01709776371717453, + "learning_rate": 0.0006, + "loss": 4.313847064971924, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.016571134328842163, + "learning_rate": 0.0006, + "loss": 4.1895856857299805, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.016851048916578293, + "learning_rate": 0.0006, + "loss": 4.350772857666016, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.017007440328598022, + "learning_rate": 0.0006, + "loss": 4.399008750915527, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.01791212148964405, + "learning_rate": 0.0006, + "loss": 4.306121826171875, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.636417865753174, + "eval_runtime": 56.8296, + "eval_samples_per_second": 42.971, + "eval_steps_per_second": 1.355, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.01729617640376091, + "learning_rate": 0.0006, + "loss": 4.257018089294434, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.017466643825173378, + "learning_rate": 0.0006, + "loss": 4.307605266571045, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.018614279106259346, + "learning_rate": 0.0006, + "loss": 4.360141277313232, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.02010771445930004, + "learning_rate": 0.0006, + "loss": 4.196428298950195, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.02090228535234928, + "learning_rate": 0.0006, + "loss": 4.288947105407715, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.02050255797803402, + "learning_rate": 0.0006, + "loss": 4.259582042694092, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.021844089031219482, + "learning_rate": 0.0006, + "loss": 4.405138969421387, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.022287018597126007, + "learning_rate": 0.0006, + "loss": 4.333793640136719, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.02288329415023327, + "learning_rate": 0.0006, + "loss": 4.3910980224609375, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.02373199723660946, + "learning_rate": 0.0006, + "loss": 4.32921028137207, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.024960234761238098, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.026743892580270767, + "learning_rate": 0.0006, + "loss": 4.315308094024658, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.022138185799121857, + "learning_rate": 0.0006, + "loss": 4.360208511352539, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.022120574489235878, + "learning_rate": 0.0006, + "loss": 4.1973772048950195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.02230502851307392, + "learning_rate": 0.0006, + "loss": 4.285558700561523, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.022301986813545227, + "learning_rate": 0.0006, + "loss": 4.238343238830566, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.020050447434186935, + "learning_rate": 0.0006, + "loss": 4.298235893249512, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.01987724006175995, + "learning_rate": 0.0006, + "loss": 4.301384449005127, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.020440855994820595, + "learning_rate": 0.0006, + "loss": 4.269218444824219, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.01891893707215786, + "learning_rate": 0.0006, + "loss": 4.2401957511901855, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.01903688907623291, + "learning_rate": 0.0006, + "loss": 4.139670372009277, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.018856938928365707, + "learning_rate": 0.0006, + "loss": 4.233214378356934, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.017791323363780975, + "learning_rate": 0.0006, + "loss": 4.2342023849487305, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.017304565757513046, + "learning_rate": 0.0006, + "loss": 4.355198860168457, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.01576206088066101, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.015105154365301132, + "learning_rate": 0.0006, + "loss": 4.313795566558838, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.01563587784767151, + "learning_rate": 0.0006, + "loss": 4.308282852172852, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.015980906784534454, + "learning_rate": 0.0006, + "loss": 4.345149040222168, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.01669284701347351, + "learning_rate": 0.0006, + "loss": 4.237611293792725, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.016202064231038094, + "learning_rate": 0.0006, + "loss": 4.3072309494018555, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.01478662807494402, + "learning_rate": 0.0006, + "loss": 4.276215076446533, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.015168731100857258, + "learning_rate": 0.0006, + "loss": 4.37509822845459, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.01648980937898159, + "learning_rate": 0.0006, + "loss": 4.349167346954346, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01708398200571537, + "learning_rate": 0.0006, + "loss": 4.346580505371094, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.017370784655213356, + "learning_rate": 0.0006, + "loss": 4.299466133117676, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.016248228028416634, + "learning_rate": 0.0006, + "loss": 4.184610366821289, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.017106913030147552, + "learning_rate": 0.0006, + "loss": 4.366856098175049, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.016071965917944908, + "learning_rate": 0.0006, + "loss": 4.297691345214844, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.016451986506581306, + "learning_rate": 0.0006, + "loss": 4.229983329772949, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.016195081174373627, + "learning_rate": 0.0006, + "loss": 4.259893417358398, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.0165175162255764, + "learning_rate": 0.0006, + "loss": 4.3348236083984375, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.01651517115533352, + "learning_rate": 0.0006, + "loss": 4.319541931152344, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.016143690794706345, + "learning_rate": 0.0006, + "loss": 4.345652103424072, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.015140696428716183, + "learning_rate": 0.0006, + "loss": 4.212047576904297, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.014637443237006664, + "learning_rate": 0.0006, + "loss": 4.187453269958496, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.014585614204406738, + "learning_rate": 0.0006, + "loss": 4.253936767578125, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.015158289112150669, + "learning_rate": 0.0006, + "loss": 4.34612512588501, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.015202849172055721, + "learning_rate": 0.0006, + "loss": 4.270634651184082, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.015363575890660286, + "learning_rate": 0.0006, + "loss": 4.271080493927002, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.014360226690769196, + "learning_rate": 0.0006, + "loss": 4.375033855438232, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.014263181947171688, + "learning_rate": 0.0006, + "loss": 4.424138069152832, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.014398688450455666, + "learning_rate": 0.0006, + "loss": 4.410154819488525, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.016131488606333733, + "learning_rate": 0.0006, + "loss": 4.324014663696289, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.017568735405802727, + "learning_rate": 0.0006, + "loss": 4.367861747741699, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.017907511442899704, + "learning_rate": 0.0006, + "loss": 4.352965354919434, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.016918016597628593, + "learning_rate": 0.0006, + "loss": 4.252397537231445, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.017274610698223114, + "learning_rate": 0.0006, + "loss": 4.342309951782227, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.017879825085401535, + "learning_rate": 0.0006, + "loss": 4.330987453460693, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.01736099272966385, + "learning_rate": 0.0006, + "loss": 4.307255744934082, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.016575824469327927, + "learning_rate": 0.0006, + "loss": 4.262016773223877, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.015181140042841434, + "learning_rate": 0.0006, + "loss": 4.291863441467285, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.015965687111020088, + "learning_rate": 0.0006, + "loss": 4.424836158752441, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.01687219925224781, + "learning_rate": 0.0006, + "loss": 4.343531131744385, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.015115504153072834, + "learning_rate": 0.0006, + "loss": 4.357808589935303, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.014150998555123806, + "learning_rate": 0.0006, + "loss": 4.324550151824951, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.014989510178565979, + "learning_rate": 0.0006, + "loss": 4.380428314208984, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.01527960691601038, + "learning_rate": 0.0006, + "loss": 4.399374961853027, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.015434633940458298, + "learning_rate": 0.0006, + "loss": 4.260984420776367, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.015370228327810764, + "learning_rate": 0.0006, + "loss": 4.3340253829956055, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.015348542481660843, + "learning_rate": 0.0006, + "loss": 4.369132995605469, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.015354936942458153, + "learning_rate": 0.0006, + "loss": 4.381937026977539, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.01796272210776806, + "learning_rate": 0.0006, + "loss": 4.286768436431885, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.576984405517578, + "eval_runtime": 57.2091, + "eval_samples_per_second": 42.685, + "eval_steps_per_second": 1.346, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.01748845726251602, + "learning_rate": 0.0006, + "loss": 4.337504863739014, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.019244296476244926, + "learning_rate": 0.0006, + "loss": 4.264280796051025, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.0215692650526762, + "learning_rate": 0.0006, + "loss": 4.350830078125, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.022189252078533173, + "learning_rate": 0.0006, + "loss": 4.299098968505859, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.022384043782949448, + "learning_rate": 0.0006, + "loss": 4.1185173988342285, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.02076014317572117, + "learning_rate": 0.0006, + "loss": 4.297612190246582, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.020843051373958588, + "learning_rate": 0.0006, + "loss": 4.294098854064941, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.020654456689953804, + "learning_rate": 0.0006, + "loss": 4.267856597900391, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.01860472559928894, + "learning_rate": 0.0006, + "loss": 4.284232139587402, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.018719421699643135, + "learning_rate": 0.0006, + "loss": 4.277214527130127, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.020903829485177994, + "learning_rate": 0.0006, + "loss": 4.254947662353516, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.02020149491727352, + "learning_rate": 0.0006, + "loss": 4.2526936531066895, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.0180392786860466, + "learning_rate": 0.0006, + "loss": 4.309922695159912, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.018163125962018967, + "learning_rate": 0.0006, + "loss": 4.13860559463501, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.01853291131556034, + "learning_rate": 0.0006, + "loss": 4.190926551818848, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.01757677271962166, + "learning_rate": 0.0006, + "loss": 4.163999557495117, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.01767992228269577, + "learning_rate": 0.0006, + "loss": 4.184430122375488, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.018139973282814026, + "learning_rate": 0.0006, + "loss": 4.300766944885254, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.020184461027383804, + "learning_rate": 0.0006, + "loss": 4.232537269592285, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.02028856985270977, + "learning_rate": 0.0006, + "loss": 4.324883460998535, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.017120616510510445, + "learning_rate": 0.0006, + "loss": 4.23493766784668, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.016202978789806366, + "learning_rate": 0.0006, + "loss": 4.197602272033691, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.017231421545147896, + "learning_rate": 0.0006, + "loss": 4.302001953125, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.01725655607879162, + "learning_rate": 0.0006, + "loss": 4.356800079345703, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.016657443717122078, + "learning_rate": 0.0006, + "loss": 4.266307830810547, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.01616556942462921, + "learning_rate": 0.0006, + "loss": 4.307868003845215, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.016621306538581848, + "learning_rate": 0.0006, + "loss": 4.172905921936035, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.016243206337094307, + "learning_rate": 0.0006, + "loss": 4.162896156311035, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.016663808375597, + "learning_rate": 0.0006, + "loss": 4.258232116699219, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.01677478663623333, + "learning_rate": 0.0006, + "loss": 4.336090087890625, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.01641303487122059, + "learning_rate": 0.0006, + "loss": 4.191476821899414, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.016700323671102524, + "learning_rate": 0.0006, + "loss": 4.299427509307861, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.017595035955309868, + "learning_rate": 0.0006, + "loss": 4.312828063964844, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.020300284028053284, + "learning_rate": 0.0006, + "loss": 4.28331184387207, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.02198321744799614, + "learning_rate": 0.0006, + "loss": 4.3280839920043945, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.02003200724720955, + "learning_rate": 0.0006, + "loss": 4.325623989105225, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.01870287023484707, + "learning_rate": 0.0006, + "loss": 4.251203536987305, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.01835448481142521, + "learning_rate": 0.0006, + "loss": 4.335144996643066, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.02046327292919159, + "learning_rate": 0.0006, + "loss": 4.238023281097412, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.019771311432123184, + "learning_rate": 0.0006, + "loss": 4.353594779968262, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.018090544268488884, + "learning_rate": 0.0006, + "loss": 4.304343223571777, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.017489777877926826, + "learning_rate": 0.0006, + "loss": 4.204375267028809, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.016785891726613045, + "learning_rate": 0.0006, + "loss": 4.286669731140137, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.016382789239287376, + "learning_rate": 0.0006, + "loss": 4.347662448883057, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.016581876203417778, + "learning_rate": 0.0006, + "loss": 4.2716193199157715, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.015627184882760048, + "learning_rate": 0.0006, + "loss": 4.311489105224609, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.015908483415842056, + "learning_rate": 0.0006, + "loss": 4.262099742889404, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.015576236881315708, + "learning_rate": 0.0006, + "loss": 4.275871276855469, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.014621064998209476, + "learning_rate": 0.0006, + "loss": 4.184260845184326, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.014797811396420002, + "learning_rate": 0.0006, + "loss": 4.3412628173828125, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.015610570088028908, + "learning_rate": 0.0006, + "loss": 4.181094646453857, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.016048265621066093, + "learning_rate": 0.0006, + "loss": 4.365290641784668, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.016604071483016014, + "learning_rate": 0.0006, + "loss": 4.2628912925720215, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.017044221982359886, + "learning_rate": 0.0006, + "loss": 4.378512382507324, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.01746384985744953, + "learning_rate": 0.0006, + "loss": 4.3888773918151855, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.01572711206972599, + "learning_rate": 0.0006, + "loss": 4.189935684204102, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.014976629987359047, + "learning_rate": 0.0006, + "loss": 4.280525207519531, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.014804664999246597, + "learning_rate": 0.0006, + "loss": 4.270630836486816, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.014641920104622841, + "learning_rate": 0.0006, + "loss": 4.319505214691162, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.014928505755960941, + "learning_rate": 0.0006, + "loss": 4.180147647857666, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.016130365431308746, + "learning_rate": 0.0006, + "loss": 4.260660648345947, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.015642493963241577, + "learning_rate": 0.0006, + "loss": 4.304064750671387, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.015988217666745186, + "learning_rate": 0.0006, + "loss": 4.2548017501831055, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.016501398757100105, + "learning_rate": 0.0006, + "loss": 4.260058879852295, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01645040698349476, + "learning_rate": 0.0006, + "loss": 4.483081340789795, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.016140863299369812, + "learning_rate": 0.0006, + "loss": 4.1776838302612305, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.016681019216775894, + "learning_rate": 0.0006, + "loss": 4.224747657775879, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.01650378853082657, + "learning_rate": 0.0006, + "loss": 4.305703163146973, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.017051683738827705, + "learning_rate": 0.0006, + "loss": 4.285835266113281, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.016894327476620674, + "learning_rate": 0.0006, + "loss": 4.318901538848877, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.017313100397586823, + "learning_rate": 0.0006, + "loss": 4.281434535980225, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.018450884148478508, + "learning_rate": 0.0006, + "loss": 4.335104942321777, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.640687465667725, + "eval_runtime": 56.9624, + "eval_samples_per_second": 42.87, + "eval_steps_per_second": 1.352, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.017755698412656784, + "learning_rate": 0.0006, + "loss": 4.278824329376221, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.01724310778081417, + "learning_rate": 0.0006, + "loss": 4.311049461364746, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.01690123789012432, + "learning_rate": 0.0006, + "loss": 4.261994361877441, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.018406696617603302, + "learning_rate": 0.0006, + "loss": 4.1925153732299805, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.019945867359638214, + "learning_rate": 0.0006, + "loss": 4.108537197113037, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.020316628739237785, + "learning_rate": 0.0006, + "loss": 4.209043025970459, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.019690370187163353, + "learning_rate": 0.0006, + "loss": 4.195789337158203, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01785232499241829, + "learning_rate": 0.0006, + "loss": 4.21466064453125, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.0164767038077116, + "learning_rate": 0.0006, + "loss": 4.270247459411621, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.018008455634117126, + "learning_rate": 0.0006, + "loss": 4.201042175292969, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.01807340793311596, + "learning_rate": 0.0006, + "loss": 4.25289249420166, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.016246909275650978, + "learning_rate": 0.0006, + "loss": 4.116283416748047, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.017545776441693306, + "learning_rate": 0.0006, + "loss": 4.316999435424805, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.017915375530719757, + "learning_rate": 0.0006, + "loss": 4.253859519958496, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.018112223595380783, + "learning_rate": 0.0006, + "loss": 4.2541022300720215, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.01846139505505562, + "learning_rate": 0.0006, + "loss": 4.20438289642334, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.01943897269666195, + "learning_rate": 0.0006, + "loss": 4.2799577713012695, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.018253348767757416, + "learning_rate": 0.0006, + "loss": 4.1720499992370605, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.018292246386408806, + "learning_rate": 0.0006, + "loss": 4.28231143951416, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.01943155936896801, + "learning_rate": 0.0006, + "loss": 4.314976692199707, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.019064994528889656, + "learning_rate": 0.0006, + "loss": 4.25328254699707, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.017423273995518684, + "learning_rate": 0.0006, + "loss": 4.195992469787598, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.01677924580872059, + "learning_rate": 0.0006, + "loss": 4.260043144226074, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.01720673404633999, + "learning_rate": 0.0006, + "loss": 4.222780704498291, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.017300743609666824, + "learning_rate": 0.0006, + "loss": 4.109524250030518, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.016065871343016624, + "learning_rate": 0.0006, + "loss": 4.36108922958374, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.016626616939902306, + "learning_rate": 0.0006, + "loss": 4.220630645751953, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.016482602804899216, + "learning_rate": 0.0006, + "loss": 4.344797134399414, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.01727953553199768, + "learning_rate": 0.0006, + "loss": 4.296645164489746, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.01522731315344572, + "learning_rate": 0.0006, + "loss": 4.223751068115234, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.01581776700913906, + "learning_rate": 0.0006, + "loss": 4.351379871368408, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.016394605860114098, + "learning_rate": 0.0006, + "loss": 4.2412919998168945, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.01751169003546238, + "learning_rate": 0.0006, + "loss": 4.257606506347656, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.017452310770750046, + "learning_rate": 0.0006, + "loss": 4.380134582519531, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.017741898074746132, + "learning_rate": 0.0006, + "loss": 4.279632568359375, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.017680590972304344, + "learning_rate": 0.0006, + "loss": 4.273406028747559, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.01864718459546566, + "learning_rate": 0.0006, + "loss": 4.278583526611328, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.019558541476726532, + "learning_rate": 0.0006, + "loss": 4.211919784545898, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.019467400386929512, + "learning_rate": 0.0006, + "loss": 4.276822566986084, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.01756172813475132, + "learning_rate": 0.0006, + "loss": 4.165379047393799, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.017458263784646988, + "learning_rate": 0.0006, + "loss": 4.2992401123046875, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.018089665099978447, + "learning_rate": 0.0006, + "loss": 4.243378639221191, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.019535524770617485, + "learning_rate": 0.0006, + "loss": 4.321477890014648, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.01997970975935459, + "learning_rate": 0.0006, + "loss": 4.123082637786865, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.01913067139685154, + "learning_rate": 0.0006, + "loss": 4.289896488189697, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.0169806070625782, + "learning_rate": 0.0006, + "loss": 4.247004985809326, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.017315007746219635, + "learning_rate": 0.0006, + "loss": 4.2095441818237305, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.015098759904503822, + "learning_rate": 0.0006, + "loss": 4.145560264587402, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.016240514814853668, + "learning_rate": 0.0006, + "loss": 4.254927635192871, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.01621108129620552, + "learning_rate": 0.0006, + "loss": 4.219725608825684, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.015191009268164635, + "learning_rate": 0.0006, + "loss": 4.194345474243164, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.014387983828783035, + "learning_rate": 0.0006, + "loss": 4.211467742919922, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.015460561029613018, + "learning_rate": 0.0006, + "loss": 4.358700752258301, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.015100076794624329, + "learning_rate": 0.0006, + "loss": 4.171223163604736, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.015607891604304314, + "learning_rate": 0.0006, + "loss": 4.217874526977539, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.015650689601898193, + "learning_rate": 0.0006, + "loss": 4.280780792236328, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.01580994389951229, + "learning_rate": 0.0006, + "loss": 4.159212112426758, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.016607413068413734, + "learning_rate": 0.0006, + "loss": 4.238853454589844, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.017383860424160957, + "learning_rate": 0.0006, + "loss": 4.233944892883301, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.014966659247875214, + "learning_rate": 0.0006, + "loss": 4.167392730712891, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.014589878730475903, + "learning_rate": 0.0006, + "loss": 4.2390899658203125, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.017179450020194054, + "learning_rate": 0.0006, + "loss": 4.365785598754883, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.01893901452422142, + "learning_rate": 0.0006, + "loss": 4.289430141448975, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.018619712442159653, + "learning_rate": 0.0006, + "loss": 4.310665130615234, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.019180385395884514, + "learning_rate": 0.0006, + "loss": 4.27454137802124, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.017811523750424385, + "learning_rate": 0.0006, + "loss": 4.312562465667725, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.016583573073148727, + "learning_rate": 0.0006, + "loss": 4.202253341674805, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.01658615842461586, + "learning_rate": 0.0006, + "loss": 4.183433532714844, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.015327401459217072, + "learning_rate": 0.0006, + "loss": 4.2125139236450195, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.015004601329565048, + "learning_rate": 0.0006, + "loss": 4.257392883300781, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.015259161591529846, + "learning_rate": 0.0006, + "loss": 4.241818428039551, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.016808776184916496, + "learning_rate": 0.0006, + "loss": 4.146241188049316, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.663090705871582, + "eval_runtime": 56.4559, + "eval_samples_per_second": 43.255, + "eval_steps_per_second": 1.364, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.015784382820129395, + "learning_rate": 0.0006, + "loss": 4.044191360473633, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.015636710450053215, + "learning_rate": 0.0006, + "loss": 4.128754615783691, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.015364975668489933, + "learning_rate": 0.0006, + "loss": 4.191835403442383, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.015015700832009315, + "learning_rate": 0.0006, + "loss": 4.120022773742676, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.016666799783706665, + "learning_rate": 0.0006, + "loss": 4.33906364440918, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.01780104823410511, + "learning_rate": 0.0006, + "loss": 4.279458999633789, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.01870882883667946, + "learning_rate": 0.0006, + "loss": 4.34728479385376, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.018406381830573082, + "learning_rate": 0.0006, + "loss": 4.211311340332031, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.017833448946475983, + "learning_rate": 0.0006, + "loss": 4.223254203796387, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.017552688717842102, + "learning_rate": 0.0006, + "loss": 4.208823204040527, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.018500229343771935, + "learning_rate": 0.0006, + "loss": 4.198025226593018, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.01782156340777874, + "learning_rate": 0.0006, + "loss": 4.179978370666504, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.019580967724323273, + "learning_rate": 0.0006, + "loss": 4.291114807128906, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.020804036408662796, + "learning_rate": 0.0006, + "loss": 4.220607280731201, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.019595693796873093, + "learning_rate": 0.0006, + "loss": 4.2905731201171875, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.019247086718678474, + "learning_rate": 0.0006, + "loss": 4.2467360496521, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.01958036608994007, + "learning_rate": 0.0006, + "loss": 4.286137580871582, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.021354753524065018, + "learning_rate": 0.0006, + "loss": 4.234755039215088, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.02206375077366829, + "learning_rate": 0.0006, + "loss": 4.149641036987305, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.02201310358941555, + "learning_rate": 0.0006, + "loss": 4.1928629875183105, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.01961950585246086, + "learning_rate": 0.0006, + "loss": 4.211404323577881, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.019211795181035995, + "learning_rate": 0.0006, + "loss": 4.200976848602295, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.01779370754957199, + "learning_rate": 0.0006, + "loss": 4.249148845672607, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.018635908141732216, + "learning_rate": 0.0006, + "loss": 4.11806058883667, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.019792694598436356, + "learning_rate": 0.0006, + "loss": 4.335224151611328, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.021731717512011528, + "learning_rate": 0.0006, + "loss": 4.302778244018555, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.020823447033762932, + "learning_rate": 0.0006, + "loss": 4.260610580444336, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.019385412335395813, + "learning_rate": 0.0006, + "loss": 4.2247724533081055, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.019539108499884605, + "learning_rate": 0.0006, + "loss": 4.236655235290527, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.01942690648138523, + "learning_rate": 0.0006, + "loss": 4.251850605010986, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.019667886197566986, + "learning_rate": 0.0006, + "loss": 4.222312927246094, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.02029012329876423, + "learning_rate": 0.0006, + "loss": 4.246252059936523, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.01784469373524189, + "learning_rate": 0.0006, + "loss": 4.192128658294678, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.0160287544131279, + "learning_rate": 0.0006, + "loss": 4.208607196807861, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.015348346903920174, + "learning_rate": 0.0006, + "loss": 4.236598014831543, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.015198206529021263, + "learning_rate": 0.0006, + "loss": 4.235552787780762, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.01523754745721817, + "learning_rate": 0.0006, + "loss": 4.1941728591918945, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.015559614636003971, + "learning_rate": 0.0006, + "loss": 4.294757843017578, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.01568358950316906, + "learning_rate": 0.0006, + "loss": 4.213065147399902, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.015336516313254833, + "learning_rate": 0.0006, + "loss": 4.240293979644775, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.015041496604681015, + "learning_rate": 0.0006, + "loss": 4.273397445678711, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.01610128954052925, + "learning_rate": 0.0006, + "loss": 4.125369071960449, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.015974976122379303, + "learning_rate": 0.0006, + "loss": 4.133459091186523, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.015936799347400665, + "learning_rate": 0.0006, + "loss": 4.306667327880859, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.01795247197151184, + "learning_rate": 0.0006, + "loss": 4.205156326293945, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.019639814272522926, + "learning_rate": 0.0006, + "loss": 4.277614593505859, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.01722300611436367, + "learning_rate": 0.0006, + "loss": 4.191032886505127, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.01612633652985096, + "learning_rate": 0.0006, + "loss": 4.20717191696167, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01603279449045658, + "learning_rate": 0.0006, + "loss": 4.228884220123291, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.016405927017331123, + "learning_rate": 0.0006, + "loss": 4.201292991638184, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.016644183546304703, + "learning_rate": 0.0006, + "loss": 4.239282131195068, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.0175587497651577, + "learning_rate": 0.0006, + "loss": 4.289038181304932, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.018223397433757782, + "learning_rate": 0.0006, + "loss": 4.2266998291015625, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.018680868670344353, + "learning_rate": 0.0006, + "loss": 4.348849296569824, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.016726728528738022, + "learning_rate": 0.0006, + "loss": 4.292080879211426, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.017386795952916145, + "learning_rate": 0.0006, + "loss": 4.208870887756348, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.01865178905427456, + "learning_rate": 0.0006, + "loss": 4.054150581359863, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.01776815392076969, + "learning_rate": 0.0006, + "loss": 4.202404499053955, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.01643497310578823, + "learning_rate": 0.0006, + "loss": 4.131855010986328, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.01574273779988289, + "learning_rate": 0.0006, + "loss": 4.296788692474365, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.015613908879458904, + "learning_rate": 0.0006, + "loss": 4.210949897766113, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.015543212182819843, + "learning_rate": 0.0006, + "loss": 4.218136310577393, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.014681251719594002, + "learning_rate": 0.0006, + "loss": 4.19401741027832, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.014393828809261322, + "learning_rate": 0.0006, + "loss": 4.198468208312988, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.01580170728266239, + "learning_rate": 0.0006, + "loss": 4.224969863891602, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.01484165620058775, + "learning_rate": 0.0006, + "loss": 4.12236213684082, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.01497623696923256, + "learning_rate": 0.0006, + "loss": 4.24141788482666, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.015080046840012074, + "learning_rate": 0.0006, + "loss": 4.284151077270508, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.016207212582230568, + "learning_rate": 0.0006, + "loss": 4.298693656921387, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.015474417246878147, + "learning_rate": 0.0006, + "loss": 4.2412261962890625, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.01515351701527834, + "learning_rate": 0.0006, + "loss": 4.3054423332214355, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.016411345452070236, + "learning_rate": 0.0006, + "loss": 4.104830741882324, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.661567687988281, + "eval_runtime": 57.1244, + "eval_samples_per_second": 42.749, + "eval_steps_per_second": 1.348, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.01665342226624489, + "learning_rate": 0.0006, + "loss": 4.311391830444336, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.018053214997053146, + "learning_rate": 0.0006, + "loss": 4.135597229003906, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.018213748931884766, + "learning_rate": 0.0006, + "loss": 4.303299903869629, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.018460242077708244, + "learning_rate": 0.0006, + "loss": 4.20850944519043, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.018365809693932533, + "learning_rate": 0.0006, + "loss": 4.275847434997559, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.018346186727285385, + "learning_rate": 0.0006, + "loss": 4.202091217041016, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.017920587211847305, + "learning_rate": 0.0006, + "loss": 4.182092666625977, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.01812003180384636, + "learning_rate": 0.0006, + "loss": 4.259721755981445, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.01700986735522747, + "learning_rate": 0.0006, + "loss": 4.067573070526123, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.018069760873913765, + "learning_rate": 0.0006, + "loss": 4.234400749206543, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.01905428245663643, + "learning_rate": 0.0006, + "loss": 4.16609001159668, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.018711242824792862, + "learning_rate": 0.0006, + "loss": 4.249088764190674, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.018721306696534157, + "learning_rate": 0.0006, + "loss": 4.14540958404541, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.020139874890446663, + "learning_rate": 0.0006, + "loss": 4.049248695373535, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.020849574357271194, + "learning_rate": 0.0006, + "loss": 4.235224723815918, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.018693549558520317, + "learning_rate": 0.0006, + "loss": 4.232961177825928, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.017889857292175293, + "learning_rate": 0.0006, + "loss": 4.171422004699707, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.016690224409103394, + "learning_rate": 0.0006, + "loss": 4.244717597961426, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.01708616502583027, + "learning_rate": 0.0006, + "loss": 4.125068664550781, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.017928237095475197, + "learning_rate": 0.0006, + "loss": 4.208292007446289, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.018250394612550735, + "learning_rate": 0.0006, + "loss": 4.118630886077881, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.020160246640443802, + "learning_rate": 0.0006, + "loss": 4.206025123596191, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.020045453682541847, + "learning_rate": 0.0006, + "loss": 4.2758684158325195, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.019556907936930656, + "learning_rate": 0.0006, + "loss": 4.188272953033447, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.019185200333595276, + "learning_rate": 0.0006, + "loss": 4.17451286315918, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.01816629432141781, + "learning_rate": 0.0006, + "loss": 4.165246963500977, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.01865146867930889, + "learning_rate": 0.0006, + "loss": 4.148184776306152, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.018338464200496674, + "learning_rate": 0.0006, + "loss": 4.288936614990234, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.018986187875270844, + "learning_rate": 0.0006, + "loss": 4.183379650115967, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.018409637734293938, + "learning_rate": 0.0006, + "loss": 4.057574272155762, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.016664542257785797, + "learning_rate": 0.0006, + "loss": 4.18517541885376, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.016486089676618576, + "learning_rate": 0.0006, + "loss": 4.179367542266846, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.01787523552775383, + "learning_rate": 0.0006, + "loss": 4.184126377105713, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.016574613749980927, + "learning_rate": 0.0006, + "loss": 4.290170192718506, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.01643376611173153, + "learning_rate": 0.0006, + "loss": 4.052913665771484, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.016503969207406044, + "learning_rate": 0.0006, + "loss": 4.335184097290039, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.017596479505300522, + "learning_rate": 0.0006, + "loss": 4.165539741516113, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.017468582838773727, + "learning_rate": 0.0006, + "loss": 4.315964698791504, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.01724618673324585, + "learning_rate": 0.0006, + "loss": 4.243681907653809, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.018384616822004318, + "learning_rate": 0.0006, + "loss": 4.112447261810303, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.01824074238538742, + "learning_rate": 0.0006, + "loss": 4.236065864562988, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.01625971868634224, + "learning_rate": 0.0006, + "loss": 4.18143892288208, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.015537494793534279, + "learning_rate": 0.0006, + "loss": 4.225147247314453, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.015818974003195763, + "learning_rate": 0.0006, + "loss": 4.3100175857543945, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.015909671783447266, + "learning_rate": 0.0006, + "loss": 4.169775009155273, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.01769658550620079, + "learning_rate": 0.0006, + "loss": 4.278904914855957, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.015641704201698303, + "learning_rate": 0.0006, + "loss": 4.2426228523254395, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.01565658301115036, + "learning_rate": 0.0006, + "loss": 4.232687950134277, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.016485940665006638, + "learning_rate": 0.0006, + "loss": 4.209839820861816, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.017622729763388634, + "learning_rate": 0.0006, + "loss": 4.196017265319824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.018320564180612564, + "learning_rate": 0.0006, + "loss": 4.225683689117432, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.018140794709324837, + "learning_rate": 0.0006, + "loss": 4.311944961547852, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.016227100044488907, + "learning_rate": 0.0006, + "loss": 4.230257511138916, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.015759488567709923, + "learning_rate": 0.0006, + "loss": 4.119174003601074, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.01757766678929329, + "learning_rate": 0.0006, + "loss": 4.165826797485352, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.019661923870444298, + "learning_rate": 0.0006, + "loss": 4.392203330993652, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.019478755071759224, + "learning_rate": 0.0006, + "loss": 4.348371982574463, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.019149569794535637, + "learning_rate": 0.0006, + "loss": 4.34104061126709, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.017707521095871925, + "learning_rate": 0.0006, + "loss": 4.228226661682129, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.015615378506481647, + "learning_rate": 0.0006, + "loss": 4.175541877746582, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.016111129894852638, + "learning_rate": 0.0006, + "loss": 4.267936706542969, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.01528779324144125, + "learning_rate": 0.0006, + "loss": 4.186014175415039, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.014925646595656872, + "learning_rate": 0.0006, + "loss": 4.174860000610352, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.016917673870921135, + "learning_rate": 0.0006, + "loss": 4.33908748626709, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.016007075086236, + "learning_rate": 0.0006, + "loss": 4.266141414642334, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.016178956255316734, + "learning_rate": 0.0006, + "loss": 4.271215915679932, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.016180871054530144, + "learning_rate": 0.0006, + "loss": 4.278801441192627, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.015335503034293652, + "learning_rate": 0.0006, + "loss": 4.194978713989258, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.016580764204263687, + "learning_rate": 0.0006, + "loss": 4.290809631347656, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.01623128168284893, + "learning_rate": 0.0006, + "loss": 4.205451488494873, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.014904248528182507, + "learning_rate": 0.0006, + "loss": 4.24924373626709, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.017089299857616425, + "learning_rate": 0.0006, + "loss": 4.222072601318359, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.585411071777344, + "eval_runtime": 57.1164, + "eval_samples_per_second": 42.755, + "eval_steps_per_second": 1.348, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.017226383090019226, + "learning_rate": 0.0006, + "loss": 4.19467830657959, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01635066606104374, + "learning_rate": 0.0006, + "loss": 4.176875591278076, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01735362410545349, + "learning_rate": 0.0006, + "loss": 4.13667106628418, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.017047109082341194, + "learning_rate": 0.0006, + "loss": 4.136329650878906, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.018116053193807602, + "learning_rate": 0.0006, + "loss": 4.228706359863281, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.01791389286518097, + "learning_rate": 0.0006, + "loss": 4.1597747802734375, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.017793502658605576, + "learning_rate": 0.0006, + "loss": 4.0807976722717285, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.01827503927052021, + "learning_rate": 0.0006, + "loss": 4.170646667480469, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.017228564247488976, + "learning_rate": 0.0006, + "loss": 4.146934509277344, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.01757437363266945, + "learning_rate": 0.0006, + "loss": 4.1859130859375, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.01701093092560768, + "learning_rate": 0.0006, + "loss": 4.126981258392334, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.015220489352941513, + "learning_rate": 0.0006, + "loss": 4.164018630981445, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.016254238784313202, + "learning_rate": 0.0006, + "loss": 4.258275508880615, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.016609614714980125, + "learning_rate": 0.0006, + "loss": 4.222233772277832, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.016531318426132202, + "learning_rate": 0.0006, + "loss": 4.144489288330078, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.01567668654024601, + "learning_rate": 0.0006, + "loss": 4.181916236877441, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.017195681110024452, + "learning_rate": 0.0006, + "loss": 4.189693927764893, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.016882948577404022, + "learning_rate": 0.0006, + "loss": 4.21604061126709, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.016261622309684753, + "learning_rate": 0.0006, + "loss": 4.227639198303223, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.016411812976002693, + "learning_rate": 0.0006, + "loss": 4.130362510681152, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.01732019893825054, + "learning_rate": 0.0006, + "loss": 4.230443000793457, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.01825905591249466, + "learning_rate": 0.0006, + "loss": 4.175817489624023, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.01887713558971882, + "learning_rate": 0.0006, + "loss": 4.169020652770996, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.019391661509871483, + "learning_rate": 0.0006, + "loss": 4.176318645477295, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.020840002223849297, + "learning_rate": 0.0006, + "loss": 4.174228668212891, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.020193178206682205, + "learning_rate": 0.0006, + "loss": 4.127919673919678, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.018449561670422554, + "learning_rate": 0.0006, + "loss": 4.229487895965576, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.01711983233690262, + "learning_rate": 0.0006, + "loss": 4.201651573181152, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.017293395474553108, + "learning_rate": 0.0006, + "loss": 4.168704032897949, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.018526358529925346, + "learning_rate": 0.0006, + "loss": 4.171864032745361, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.019287291914224625, + "learning_rate": 0.0006, + "loss": 4.039534568786621, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.016902444884181023, + "learning_rate": 0.0006, + "loss": 4.282787799835205, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.01656750962138176, + "learning_rate": 0.0006, + "loss": 4.213098526000977, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.016805065795779228, + "learning_rate": 0.0006, + "loss": 4.1583476066589355, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.017166638746857643, + "learning_rate": 0.0006, + "loss": 4.188634395599365, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.01592106744647026, + "learning_rate": 0.0006, + "loss": 4.234043121337891, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.015270989388227463, + "learning_rate": 0.0006, + "loss": 4.301420211791992, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.015194724313914776, + "learning_rate": 0.0006, + "loss": 4.18703031539917, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.015587746165692806, + "learning_rate": 0.0006, + "loss": 4.138928413391113, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.016029570251703262, + "learning_rate": 0.0006, + "loss": 4.290863990783691, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.015651309862732887, + "learning_rate": 0.0006, + "loss": 4.272062301635742, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.015992306172847748, + "learning_rate": 0.0006, + "loss": 4.2487406730651855, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.014969157986342907, + "learning_rate": 0.0006, + "loss": 4.115347862243652, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.014639027416706085, + "learning_rate": 0.0006, + "loss": 4.2207489013671875, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.015059413388371468, + "learning_rate": 0.0006, + "loss": 4.274044990539551, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.0166641678661108, + "learning_rate": 0.0006, + "loss": 4.191373348236084, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.017806345596909523, + "learning_rate": 0.0006, + "loss": 4.139800548553467, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.018712421879172325, + "learning_rate": 0.0006, + "loss": 4.035965442657471, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.018457243219017982, + "learning_rate": 0.0006, + "loss": 4.2579240798950195, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.01920999586582184, + "learning_rate": 0.0006, + "loss": 4.131112098693848, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.01729944534599781, + "learning_rate": 0.0006, + "loss": 4.254059314727783, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.016914231702685356, + "learning_rate": 0.0006, + "loss": 4.233428955078125, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.01913098618388176, + "learning_rate": 0.0006, + "loss": 4.23525333404541, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.019150281324982643, + "learning_rate": 0.0006, + "loss": 4.213695049285889, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.018947452306747437, + "learning_rate": 0.0006, + "loss": 4.151697158813477, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.016058508306741714, + "learning_rate": 0.0006, + "loss": 4.220280170440674, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.017488619312644005, + "learning_rate": 0.0006, + "loss": 4.154428482055664, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.017355090007185936, + "learning_rate": 0.0006, + "loss": 4.256086349487305, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.016636032611131668, + "learning_rate": 0.0006, + "loss": 4.227564334869385, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.016273891553282738, + "learning_rate": 0.0006, + "loss": 4.200466156005859, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.016939911991357803, + "learning_rate": 0.0006, + "loss": 4.1571221351623535, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.018332522362470627, + "learning_rate": 0.0006, + "loss": 4.228398323059082, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.019508114084601402, + "learning_rate": 0.0006, + "loss": 4.222960472106934, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.018440278246998787, + "learning_rate": 0.0006, + "loss": 4.261051177978516, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.018113229423761368, + "learning_rate": 0.0006, + "loss": 4.109235763549805, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.015392083674669266, + "learning_rate": 0.0006, + "loss": 4.21950101852417, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.01633290946483612, + "learning_rate": 0.0006, + "loss": 4.176053047180176, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.016236383467912674, + "learning_rate": 0.0006, + "loss": 4.314309120178223, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.015284373424947262, + "learning_rate": 0.0006, + "loss": 4.303133487701416, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.015982897952198982, + "learning_rate": 0.0006, + "loss": 4.25739860534668, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.016682198271155357, + "learning_rate": 0.0006, + "loss": 4.292210578918457, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.017521467059850693, + "learning_rate": 0.0006, + "loss": 4.170637130737305, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.642786979675293, + "eval_runtime": 57.029, + "eval_samples_per_second": 42.82, + "eval_steps_per_second": 1.35, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.016607852652668953, + "learning_rate": 0.0006, + "loss": 4.106224536895752, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.018448904156684875, + "learning_rate": 0.0006, + "loss": 4.224459648132324, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.020481228828430176, + "learning_rate": 0.0006, + "loss": 4.305548667907715, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.01999707892537117, + "learning_rate": 0.0006, + "loss": 4.253663539886475, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.020428303629159927, + "learning_rate": 0.0006, + "loss": 4.166954040527344, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.022608213126659393, + "learning_rate": 0.0006, + "loss": 4.172819137573242, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.024669520556926727, + "learning_rate": 0.0006, + "loss": 4.106586456298828, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.02247351035475731, + "learning_rate": 0.0006, + "loss": 4.160271644592285, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.0207810141146183, + "learning_rate": 0.0006, + "loss": 4.272491931915283, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.021576499566435814, + "learning_rate": 0.0006, + "loss": 4.171474456787109, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.02337106689810753, + "learning_rate": 0.0006, + "loss": 4.094712734222412, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.02219875156879425, + "learning_rate": 0.0006, + "loss": 4.1800994873046875, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.022680504247546196, + "learning_rate": 0.0006, + "loss": 4.2054362297058105, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.0208734143525362, + "learning_rate": 0.0006, + "loss": 4.06068229675293, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.01811066083610058, + "learning_rate": 0.0006, + "loss": 4.2465128898620605, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.018571637570858, + "learning_rate": 0.0006, + "loss": 4.191955089569092, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.018598807975649834, + "learning_rate": 0.0006, + "loss": 4.200872421264648, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.017577625811100006, + "learning_rate": 0.0006, + "loss": 4.148700714111328, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.01817980408668518, + "learning_rate": 0.0006, + "loss": 4.256978988647461, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.017983486875891685, + "learning_rate": 0.0006, + "loss": 4.175005912780762, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.016751961782574654, + "learning_rate": 0.0006, + "loss": 4.237518310546875, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.016407202929258347, + "learning_rate": 0.0006, + "loss": 4.197269439697266, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.01578645594418049, + "learning_rate": 0.0006, + "loss": 4.179751396179199, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.016001226380467415, + "learning_rate": 0.0006, + "loss": 4.08821964263916, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.015122811309993267, + "learning_rate": 0.0006, + "loss": 4.047828674316406, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.015399965457618237, + "learning_rate": 0.0006, + "loss": 4.1506524085998535, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.01534605398774147, + "learning_rate": 0.0006, + "loss": 4.27952766418457, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.015461381524801254, + "learning_rate": 0.0006, + "loss": 4.27102518081665, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.016769153997302055, + "learning_rate": 0.0006, + "loss": 4.191592216491699, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.016557959839701653, + "learning_rate": 0.0006, + "loss": 4.159705638885498, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.01659569889307022, + "learning_rate": 0.0006, + "loss": 4.2136664390563965, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.017101937904953957, + "learning_rate": 0.0006, + "loss": 4.2716498374938965, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.01722985878586769, + "learning_rate": 0.0006, + "loss": 4.300534248352051, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.016800185665488243, + "learning_rate": 0.0006, + "loss": 4.113240718841553, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.01603875495493412, + "learning_rate": 0.0006, + "loss": 4.157105445861816, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.015903932973742485, + "learning_rate": 0.0006, + "loss": 4.233669757843018, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.0160983856767416, + "learning_rate": 0.0006, + "loss": 4.147704124450684, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.01604730263352394, + "learning_rate": 0.0006, + "loss": 4.1465911865234375, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.016205785796046257, + "learning_rate": 0.0006, + "loss": 4.106704235076904, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.017008036375045776, + "learning_rate": 0.0006, + "loss": 4.191000938415527, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.01699635200202465, + "learning_rate": 0.0006, + "loss": 4.187042236328125, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.01692967116832733, + "learning_rate": 0.0006, + "loss": 4.208805561065674, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.017608756199479103, + "learning_rate": 0.0006, + "loss": 4.200512886047363, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.016590990126132965, + "learning_rate": 0.0006, + "loss": 4.194705009460449, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.0170235987752676, + "learning_rate": 0.0006, + "loss": 4.210326194763184, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.01604801043868065, + "learning_rate": 0.0006, + "loss": 4.196024417877197, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.015694202855229378, + "learning_rate": 0.0006, + "loss": 4.178236484527588, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.015473959036171436, + "learning_rate": 0.0006, + "loss": 4.259339809417725, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.014689362607896328, + "learning_rate": 0.0006, + "loss": 3.9702324867248535, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.015245123766362667, + "learning_rate": 0.0006, + "loss": 4.197025775909424, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.016148347407579422, + "learning_rate": 0.0006, + "loss": 4.231856346130371, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.015169818885624409, + "learning_rate": 0.0006, + "loss": 4.230169296264648, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.014843013137578964, + "learning_rate": 0.0006, + "loss": 4.195633411407471, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.015229340642690659, + "learning_rate": 0.0006, + "loss": 4.231280326843262, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.01551708485931158, + "learning_rate": 0.0006, + "loss": 4.17064094543457, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.015863575041294098, + "learning_rate": 0.0006, + "loss": 4.1579484939575195, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.015118198469281197, + "learning_rate": 0.0006, + "loss": 4.1626787185668945, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.015220776200294495, + "learning_rate": 0.0006, + "loss": 4.135746955871582, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.014529784210026264, + "learning_rate": 0.0006, + "loss": 4.164931297302246, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.015533296391367912, + "learning_rate": 0.0006, + "loss": 4.112464904785156, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.016562335193157196, + "learning_rate": 0.0006, + "loss": 4.164898872375488, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.016258113086223602, + "learning_rate": 0.0006, + "loss": 4.1979265213012695, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.01558110024780035, + "learning_rate": 0.0006, + "loss": 4.174675941467285, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.016412515193223953, + "learning_rate": 0.0006, + "loss": 4.217909336090088, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.015118442475795746, + "learning_rate": 0.0006, + "loss": 4.137143611907959, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.01506161317229271, + "learning_rate": 0.0006, + "loss": 4.196122169494629, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.017165271565318108, + "learning_rate": 0.0006, + "loss": 4.14834451675415, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.01696440577507019, + "learning_rate": 0.0006, + "loss": 4.218868732452393, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.017349394038319588, + "learning_rate": 0.0006, + "loss": 4.2754011154174805, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.018387921154499054, + "learning_rate": 0.0006, + "loss": 4.282623291015625, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.018837926909327507, + "learning_rate": 0.0006, + "loss": 4.185425281524658, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.018998464569449425, + "learning_rate": 0.0006, + "loss": 4.151291847229004, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.522614002227783, + "eval_runtime": 57.0106, + "eval_samples_per_second": 42.834, + "eval_steps_per_second": 1.351, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.019036833196878433, + "learning_rate": 0.0006, + "loss": 4.183202266693115, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.02207397110760212, + "learning_rate": 0.0006, + "loss": 4.17875862121582, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.022008279338479042, + "learning_rate": 0.0006, + "loss": 4.08042049407959, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.022407017648220062, + "learning_rate": 0.0006, + "loss": 4.112189769744873, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.021599747240543365, + "learning_rate": 0.0006, + "loss": 4.126455307006836, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.020340966060757637, + "learning_rate": 0.0006, + "loss": 4.080538272857666, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.020956892520189285, + "learning_rate": 0.0006, + "loss": 4.171943664550781, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.019966667518019676, + "learning_rate": 0.0006, + "loss": 4.144007682800293, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.018256952986121178, + "learning_rate": 0.0006, + "loss": 4.208107948303223, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.017289170995354652, + "learning_rate": 0.0006, + "loss": 4.182024955749512, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.017656700685620308, + "learning_rate": 0.0006, + "loss": 4.068373680114746, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.017903361469507217, + "learning_rate": 0.0006, + "loss": 4.2345428466796875, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.01785109005868435, + "learning_rate": 0.0006, + "loss": 4.219776153564453, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.01887257769703865, + "learning_rate": 0.0006, + "loss": 4.139206409454346, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.01924685761332512, + "learning_rate": 0.0006, + "loss": 4.121052265167236, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.020574072375893593, + "learning_rate": 0.0006, + "loss": 4.13262414932251, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.023849064484238625, + "learning_rate": 0.0006, + "loss": 4.239119529724121, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.026239361613988876, + "learning_rate": 0.0006, + "loss": 4.248993873596191, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.023980243131518364, + "learning_rate": 0.0006, + "loss": 4.159852504730225, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.020442405715584755, + "learning_rate": 0.0006, + "loss": 4.133546829223633, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.020661186426877975, + "learning_rate": 0.0006, + "loss": 4.078243732452393, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.02046387456357479, + "learning_rate": 0.0006, + "loss": 4.172012805938721, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.02089597098529339, + "learning_rate": 0.0006, + "loss": 4.0604705810546875, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.01875591278076172, + "learning_rate": 0.0006, + "loss": 4.125911235809326, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.018606796860694885, + "learning_rate": 0.0006, + "loss": 4.097362518310547, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.01766626164317131, + "learning_rate": 0.0006, + "loss": 4.118616580963135, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.017692960798740387, + "learning_rate": 0.0006, + "loss": 4.1457672119140625, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.017619704827666283, + "learning_rate": 0.0006, + "loss": 4.175961494445801, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.01784413494169712, + "learning_rate": 0.0006, + "loss": 4.051291465759277, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.018508171662688255, + "learning_rate": 0.0006, + "loss": 4.0638556480407715, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.01731436885893345, + "learning_rate": 0.0006, + "loss": 4.1805419921875, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.015886498615145683, + "learning_rate": 0.0006, + "loss": 4.151026725769043, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.016583820804953575, + "learning_rate": 0.0006, + "loss": 4.203604698181152, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.0176097322255373, + "learning_rate": 0.0006, + "loss": 4.135159015655518, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.016657903790473938, + "learning_rate": 0.0006, + "loss": 4.126681804656982, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.016866130754351616, + "learning_rate": 0.0006, + "loss": 4.162341117858887, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.016286680474877357, + "learning_rate": 0.0006, + "loss": 4.261801242828369, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.01481365505605936, + "learning_rate": 0.0006, + "loss": 4.110016822814941, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.016605013981461525, + "learning_rate": 0.0006, + "loss": 4.230024814605713, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.01645175740122795, + "learning_rate": 0.0006, + "loss": 4.174612045288086, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.014725026674568653, + "learning_rate": 0.0006, + "loss": 4.066371440887451, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.015003837645053864, + "learning_rate": 0.0006, + "loss": 4.193343639373779, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.015346302650868893, + "learning_rate": 0.0006, + "loss": 4.210659980773926, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.015135962516069412, + "learning_rate": 0.0006, + "loss": 4.204618453979492, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.01601956970989704, + "learning_rate": 0.0006, + "loss": 4.145564556121826, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.01517851185053587, + "learning_rate": 0.0006, + "loss": 4.051673889160156, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.014894292689859867, + "learning_rate": 0.0006, + "loss": 4.184818267822266, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.015720834955573082, + "learning_rate": 0.0006, + "loss": 4.220036506652832, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.01573433168232441, + "learning_rate": 0.0006, + "loss": 4.076773166656494, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.015828052535653114, + "learning_rate": 0.0006, + "loss": 4.116329669952393, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.015462259761989117, + "learning_rate": 0.0006, + "loss": 4.1155242919921875, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.015763849020004272, + "learning_rate": 0.0006, + "loss": 4.169687747955322, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.014935145154595375, + "learning_rate": 0.0006, + "loss": 4.2090911865234375, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.015119184739887714, + "learning_rate": 0.0006, + "loss": 4.131838798522949, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.014833525754511356, + "learning_rate": 0.0006, + "loss": 4.17618465423584, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.01588359661400318, + "learning_rate": 0.0006, + "loss": 4.163723945617676, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.0168188214302063, + "learning_rate": 0.0006, + "loss": 4.178189277648926, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.01726095750927925, + "learning_rate": 0.0006, + "loss": 4.144223690032959, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.01701410301029682, + "learning_rate": 0.0006, + "loss": 4.209896087646484, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.01682264916598797, + "learning_rate": 0.0006, + "loss": 4.26223087310791, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.015597456134855747, + "learning_rate": 0.0006, + "loss": 4.292664051055908, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.015258579514920712, + "learning_rate": 0.0006, + "loss": 4.165006637573242, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.01662200503051281, + "learning_rate": 0.0006, + "loss": 4.309786319732666, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.015691161155700684, + "learning_rate": 0.0006, + "loss": 4.167202949523926, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.014870637096464634, + "learning_rate": 0.0006, + "loss": 4.142415523529053, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.01567252166569233, + "learning_rate": 0.0006, + "loss": 4.2312092781066895, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.014948616735637188, + "learning_rate": 0.0006, + "loss": 4.1514129638671875, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.014654500409960747, + "learning_rate": 0.0006, + "loss": 4.205010890960693, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.015059148892760277, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.015026576817035675, + "learning_rate": 0.0006, + "loss": 4.182535171508789, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.016394320875406265, + "learning_rate": 0.0006, + "loss": 4.129051208496094, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01867043413221836, + "learning_rate": 0.0006, + "loss": 4.185556411743164, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.640405654907227, + "eval_runtime": 56.5861, + "eval_samples_per_second": 43.155, + "eval_steps_per_second": 1.361, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.018626727163791656, + "learning_rate": 0.0006, + "loss": 4.143031597137451, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.019283363595604897, + "learning_rate": 0.0006, + "loss": 4.117845058441162, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.018185822293162346, + "learning_rate": 0.0006, + "loss": 4.076862812042236, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.019411196932196617, + "learning_rate": 0.0006, + "loss": 4.124156475067139, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.019248466938734055, + "learning_rate": 0.0006, + "loss": 4.201735019683838, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.018856720998883247, + "learning_rate": 0.0006, + "loss": 4.00296688079834, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.018554294481873512, + "learning_rate": 0.0006, + "loss": 4.270577430725098, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.019946224987506866, + "learning_rate": 0.0006, + "loss": 4.119589328765869, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.021395670250058174, + "learning_rate": 0.0006, + "loss": 4.128847122192383, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.02009674906730652, + "learning_rate": 0.0006, + "loss": 4.183228969573975, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01820223033428192, + "learning_rate": 0.0006, + "loss": 4.131601810455322, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.018653156235814095, + "learning_rate": 0.0006, + "loss": 4.116430282592773, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.016828063875436783, + "learning_rate": 0.0006, + "loss": 4.054663181304932, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.110044956207275, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.016608452424407005, + "learning_rate": 0.0006, + "loss": 4.102668285369873, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.01743577979505062, + "learning_rate": 0.0006, + "loss": 4.166791915893555, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.017253734171390533, + "learning_rate": 0.0006, + "loss": 4.217832088470459, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.017937595024704933, + "learning_rate": 0.0006, + "loss": 4.2173967361450195, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.01985993981361389, + "learning_rate": 0.0006, + "loss": 4.17849588394165, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.019243355840444565, + "learning_rate": 0.0006, + "loss": 4.086523056030273, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.01924416422843933, + "learning_rate": 0.0006, + "loss": 4.128574371337891, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.01983758620917797, + "learning_rate": 0.0006, + "loss": 4.061581611633301, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.01930462196469307, + "learning_rate": 0.0006, + "loss": 4.091731071472168, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.017988424748182297, + "learning_rate": 0.0006, + "loss": 4.155599117279053, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.015194767154753208, + "learning_rate": 0.0006, + "loss": 4.144134044647217, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.01536081824451685, + "learning_rate": 0.0006, + "loss": 4.127272129058838, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.015431297942996025, + "learning_rate": 0.0006, + "loss": 4.3213019371032715, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.015740280970931053, + "learning_rate": 0.0006, + "loss": 4.121467590332031, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.015379384160041809, + "learning_rate": 0.0006, + "loss": 4.095216751098633, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.01528073474764824, + "learning_rate": 0.0006, + "loss": 4.13047981262207, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.015739979222416878, + "learning_rate": 0.0006, + "loss": 4.2533135414123535, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.01519764307886362, + "learning_rate": 0.0006, + "loss": 4.131782054901123, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.015425996854901314, + "learning_rate": 0.0006, + "loss": 4.02811336517334, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.015378830023109913, + "learning_rate": 0.0006, + "loss": 4.08652925491333, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.01473075058311224, + "learning_rate": 0.0006, + "loss": 4.214364528656006, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.015141637064516544, + "learning_rate": 0.0006, + "loss": 4.155712127685547, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.01641593500971794, + "learning_rate": 0.0006, + "loss": 4.097907066345215, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.015182029455900192, + "learning_rate": 0.0006, + "loss": 4.238499641418457, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.01643732376396656, + "learning_rate": 0.0006, + "loss": 4.0661492347717285, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.015614228323101997, + "learning_rate": 0.0006, + "loss": 4.120795249938965, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.016066187992691994, + "learning_rate": 0.0006, + "loss": 4.096465110778809, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.017214342951774597, + "learning_rate": 0.0006, + "loss": 4.242628574371338, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.01787586510181427, + "learning_rate": 0.0006, + "loss": 4.149589538574219, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.019455241039395332, + "learning_rate": 0.0006, + "loss": 4.267796039581299, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.01984674669802189, + "learning_rate": 0.0006, + "loss": 4.175220966339111, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.021933717653155327, + "learning_rate": 0.0006, + "loss": 4.106647491455078, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.02123589813709259, + "learning_rate": 0.0006, + "loss": 4.204996109008789, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.019814589992165565, + "learning_rate": 0.0006, + "loss": 4.186171531677246, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.02059728465974331, + "learning_rate": 0.0006, + "loss": 4.170509338378906, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.018483608961105347, + "learning_rate": 0.0006, + "loss": 4.085585594177246, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.018403852358460426, + "learning_rate": 0.0006, + "loss": 4.2551774978637695, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.018653830513358116, + "learning_rate": 0.0006, + "loss": 4.091766357421875, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.017692334949970245, + "learning_rate": 0.0006, + "loss": 4.100838661193848, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.016029126942157745, + "learning_rate": 0.0006, + "loss": 4.1530680656433105, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.016095684841275215, + "learning_rate": 0.0006, + "loss": 4.20544958114624, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.016375798732042313, + "learning_rate": 0.0006, + "loss": 4.199177265167236, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.01590786688029766, + "learning_rate": 0.0006, + "loss": 4.200799465179443, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.014896097593009472, + "learning_rate": 0.0006, + "loss": 4.185850143432617, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.01438513956964016, + "learning_rate": 0.0006, + "loss": 4.214962005615234, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.014459339901804924, + "learning_rate": 0.0006, + "loss": 4.196972370147705, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.014782169833779335, + "learning_rate": 0.0006, + "loss": 4.0885910987854, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.01520370040088892, + "learning_rate": 0.0006, + "loss": 4.0323262214660645, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.014056282117962837, + "learning_rate": 0.0006, + "loss": 4.1386518478393555, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.013567990623414516, + "learning_rate": 0.0006, + "loss": 4.095754623413086, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.013306716457009315, + "learning_rate": 0.0006, + "loss": 4.1623077392578125, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.014047150500118732, + "learning_rate": 0.0006, + "loss": 4.142541885375977, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.01377617847174406, + "learning_rate": 0.0006, + "loss": 4.268551826477051, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.014671621844172478, + "learning_rate": 0.0006, + "loss": 4.198196887969971, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.014806744642555714, + "learning_rate": 0.0006, + "loss": 4.057899475097656, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.015734272077679634, + "learning_rate": 0.0006, + "loss": 4.163505554199219, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.015250146389007568, + "learning_rate": 0.0006, + "loss": 4.076294898986816, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.016733702272176743, + "learning_rate": 0.0006, + "loss": 4.138420104980469, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.542503356933594, + "eval_runtime": 56.983, + "eval_samples_per_second": 42.855, + "eval_steps_per_second": 1.351, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.015866613015532494, + "learning_rate": 0.0006, + "loss": 4.032081127166748, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01946570910513401, + "learning_rate": 0.0006, + "loss": 4.122320175170898, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.020265938714146614, + "learning_rate": 0.0006, + "loss": 4.06460428237915, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.018887171521782875, + "learning_rate": 0.0006, + "loss": 4.096906661987305, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.017304692417383194, + "learning_rate": 0.0006, + "loss": 4.137261390686035, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.017029082402586937, + "learning_rate": 0.0006, + "loss": 4.081112861633301, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.016414813697338104, + "learning_rate": 0.0006, + "loss": 4.140783309936523, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.018238281831145287, + "learning_rate": 0.0006, + "loss": 4.161116600036621, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.019929232075810432, + "learning_rate": 0.0006, + "loss": 4.243722438812256, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.019375786185264587, + "learning_rate": 0.0006, + "loss": 4.134873390197754, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.018151137977838516, + "learning_rate": 0.0006, + "loss": 4.098745346069336, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.017054934054613113, + "learning_rate": 0.0006, + "loss": 4.194662570953369, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.018220555037260056, + "learning_rate": 0.0006, + "loss": 4.128866195678711, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.01692879945039749, + "learning_rate": 0.0006, + "loss": 4.062477111816406, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.01674199476838112, + "learning_rate": 0.0006, + "loss": 4.186149597167969, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.016527455300092697, + "learning_rate": 0.0006, + "loss": 4.0734028816223145, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.016448386013507843, + "learning_rate": 0.0006, + "loss": 4.098896026611328, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.014644953422248363, + "learning_rate": 0.0006, + "loss": 4.1031904220581055, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.016649341210722923, + "learning_rate": 0.0006, + "loss": 4.148791313171387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.017442714422941208, + "learning_rate": 0.0006, + "loss": 4.070799827575684, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.017852013930678368, + "learning_rate": 0.0006, + "loss": 4.031924247741699, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.016371270641684532, + "learning_rate": 0.0006, + "loss": 4.149420738220215, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.015846814960241318, + "learning_rate": 0.0006, + "loss": 4.15427303314209, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.017411163076758385, + "learning_rate": 0.0006, + "loss": 4.1304192543029785, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.01780046708881855, + "learning_rate": 0.0006, + "loss": 4.191732406616211, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.01958383060991764, + "learning_rate": 0.0006, + "loss": 4.068355083465576, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.022173317149281502, + "learning_rate": 0.0006, + "loss": 4.182929039001465, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.022389156743884087, + "learning_rate": 0.0006, + "loss": 4.110218048095703, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.022478511556982994, + "learning_rate": 0.0006, + "loss": 4.095479488372803, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.019504647701978683, + "learning_rate": 0.0006, + "loss": 4.101964950561523, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.019378576427698135, + "learning_rate": 0.0006, + "loss": 4.14620304107666, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.017993204295635223, + "learning_rate": 0.0006, + "loss": 4.183492660522461, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.01814805157482624, + "learning_rate": 0.0006, + "loss": 4.151247978210449, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.018622448667883873, + "learning_rate": 0.0006, + "loss": 4.020938396453857, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.018360132351517677, + "learning_rate": 0.0006, + "loss": 4.064734935760498, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.017715711146593094, + "learning_rate": 0.0006, + "loss": 4.185687065124512, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.01714295893907547, + "learning_rate": 0.0006, + "loss": 4.1044721603393555, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.016943803057074547, + "learning_rate": 0.0006, + "loss": 4.106837272644043, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.0182423684746027, + "learning_rate": 0.0006, + "loss": 4.1041340827941895, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.01636587455868721, + "learning_rate": 0.0006, + "loss": 3.9735546112060547, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.014593546278774738, + "learning_rate": 0.0006, + "loss": 4.041980266571045, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.015436965972185135, + "learning_rate": 0.0006, + "loss": 4.301148414611816, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.015075269155204296, + "learning_rate": 0.0006, + "loss": 4.118527412414551, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.01701514609158039, + "learning_rate": 0.0006, + "loss": 4.245968818664551, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.018047887831926346, + "learning_rate": 0.0006, + "loss": 4.227261066436768, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.018499085679650307, + "learning_rate": 0.0006, + "loss": 4.121360778808594, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.018987711519002914, + "learning_rate": 0.0006, + "loss": 4.131564617156982, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.01834476925432682, + "learning_rate": 0.0006, + "loss": 4.224941253662109, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.017507942393422127, + "learning_rate": 0.0006, + "loss": 4.1394548416137695, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.01680580899119377, + "learning_rate": 0.0006, + "loss": 4.1073431968688965, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.015233514830470085, + "learning_rate": 0.0006, + "loss": 4.136994361877441, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.015482917428016663, + "learning_rate": 0.0006, + "loss": 4.110719680786133, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.015940461307764053, + "learning_rate": 0.0006, + "loss": 4.207851886749268, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.015355597250163555, + "learning_rate": 0.0006, + "loss": 4.169469833374023, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.015974000096321106, + "learning_rate": 0.0006, + "loss": 4.110061168670654, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.016088927164673805, + "learning_rate": 0.0006, + "loss": 4.155452728271484, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.016156120225787163, + "learning_rate": 0.0006, + "loss": 4.178092956542969, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.014928647316992283, + "learning_rate": 0.0006, + "loss": 4.05457878112793, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.015289328061044216, + "learning_rate": 0.0006, + "loss": 4.18869686126709, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.015799039974808693, + "learning_rate": 0.0006, + "loss": 4.106073379516602, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.015824418514966965, + "learning_rate": 0.0006, + "loss": 4.0634589195251465, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.015321656130254269, + "learning_rate": 0.0006, + "loss": 4.093090057373047, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.014380166307091713, + "learning_rate": 0.0006, + "loss": 4.1750640869140625, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.014962567947804928, + "learning_rate": 0.0006, + "loss": 4.185277938842773, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.014467543922364712, + "learning_rate": 0.0006, + "loss": 4.1600799560546875, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.01545181218534708, + "learning_rate": 0.0006, + "loss": 4.096299648284912, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.01605340465903282, + "learning_rate": 0.0006, + "loss": 4.203323841094971, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.015203815884888172, + "learning_rate": 0.0006, + "loss": 4.201385021209717, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.015149409882724285, + "learning_rate": 0.0006, + "loss": 4.088164329528809, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.014885558746755123, + "learning_rate": 0.0006, + "loss": 4.073911666870117, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.015003419481217861, + "learning_rate": 0.0006, + "loss": 4.157113552093506, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.01641916297376156, + "learning_rate": 0.0006, + "loss": 3.970052719116211, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.641692161560059, + "eval_runtime": 56.8836, + "eval_samples_per_second": 42.93, + "eval_steps_per_second": 1.354, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.015894673764705658, + "learning_rate": 0.0006, + "loss": 4.104437351226807, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.0173867866396904, + "learning_rate": 0.0006, + "loss": 4.091726303100586, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.019272832199931145, + "learning_rate": 0.0006, + "loss": 4.217905044555664, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.021820371970534325, + "learning_rate": 0.0006, + "loss": 4.111428260803223, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.02340741828083992, + "learning_rate": 0.0006, + "loss": 4.115771293640137, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.023760691285133362, + "learning_rate": 0.0006, + "loss": 4.132383346557617, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.02073468267917633, + "learning_rate": 0.0006, + "loss": 4.272691249847412, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.02197718247771263, + "learning_rate": 0.0006, + "loss": 4.105586528778076, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.0200254675000906, + "learning_rate": 0.0006, + "loss": 3.968820095062256, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.01929587684571743, + "learning_rate": 0.0006, + "loss": 4.045468807220459, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.01886162906885147, + "learning_rate": 0.0006, + "loss": 4.185571670532227, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.019789839163422585, + "learning_rate": 0.0006, + "loss": 3.9734551906585693, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.0198142658919096, + "learning_rate": 0.0006, + "loss": 4.2662506103515625, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.02087012305855751, + "learning_rate": 0.0006, + "loss": 4.175931930541992, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.020294535905122757, + "learning_rate": 0.0006, + "loss": 4.060324668884277, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.01908983662724495, + "learning_rate": 0.0006, + "loss": 4.217939853668213, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.0178892333060503, + "learning_rate": 0.0006, + "loss": 4.190216541290283, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.017892315983772278, + "learning_rate": 0.0006, + "loss": 4.067419052124023, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.018556250259280205, + "learning_rate": 0.0006, + "loss": 4.105958461761475, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.01944422349333763, + "learning_rate": 0.0006, + "loss": 4.155775547027588, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.017333146184682846, + "learning_rate": 0.0006, + "loss": 4.244513511657715, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.01639244332909584, + "learning_rate": 0.0006, + "loss": 4.167815685272217, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.01610143668949604, + "learning_rate": 0.0006, + "loss": 4.121486663818359, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.01656411588191986, + "learning_rate": 0.0006, + "loss": 4.1755499839782715, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.015075297094881535, + "learning_rate": 0.0006, + "loss": 4.102992534637451, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.01538170501589775, + "learning_rate": 0.0006, + "loss": 4.161603927612305, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.01483312901109457, + "learning_rate": 0.0006, + "loss": 4.085355758666992, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.014894738793373108, + "learning_rate": 0.0006, + "loss": 4.083585262298584, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.01475780550390482, + "learning_rate": 0.0006, + "loss": 4.012645721435547, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.014531653374433517, + "learning_rate": 0.0006, + "loss": 4.197113990783691, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.015316493809223175, + "learning_rate": 0.0006, + "loss": 4.227970123291016, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.014685571193695068, + "learning_rate": 0.0006, + "loss": 4.089415550231934, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.015337035991251469, + "learning_rate": 0.0006, + "loss": 4.080583572387695, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.016754208132624626, + "learning_rate": 0.0006, + "loss": 4.163935661315918, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.017535798251628876, + "learning_rate": 0.0006, + "loss": 4.093027591705322, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.017336975783109665, + "learning_rate": 0.0006, + "loss": 4.127961158752441, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.016242152079939842, + "learning_rate": 0.0006, + "loss": 4.155593395233154, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.015121322125196457, + "learning_rate": 0.0006, + "loss": 3.984004020690918, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.015178884379565716, + "learning_rate": 0.0006, + "loss": 4.111564636230469, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.014777131378650665, + "learning_rate": 0.0006, + "loss": 4.041402816772461, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.015197676606476307, + "learning_rate": 0.0006, + "loss": 4.133706092834473, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.015905972570180893, + "learning_rate": 0.0006, + "loss": 4.1069135665893555, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.015942877158522606, + "learning_rate": 0.0006, + "loss": 4.0720624923706055, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.016697583720088005, + "learning_rate": 0.0006, + "loss": 4.167448043823242, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.015603577718138695, + "learning_rate": 0.0006, + "loss": 4.127327919006348, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.017153460532426834, + "learning_rate": 0.0006, + "loss": 4.056543350219727, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.016731958836317062, + "learning_rate": 0.0006, + "loss": 4.0681843757629395, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.016167359426617622, + "learning_rate": 0.0006, + "loss": 4.174992084503174, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.01615108922123909, + "learning_rate": 0.0006, + "loss": 4.1383161544799805, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.015750816091895103, + "learning_rate": 0.0006, + "loss": 4.031595706939697, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.015416189096868038, + "learning_rate": 0.0006, + "loss": 4.067498207092285, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.01570792682468891, + "learning_rate": 0.0006, + "loss": 4.1338300704956055, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.016367752104997635, + "learning_rate": 0.0006, + "loss": 4.152523040771484, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.016349803656339645, + "learning_rate": 0.0006, + "loss": 4.101985454559326, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.01663660630583763, + "learning_rate": 0.0006, + "loss": 4.190130710601807, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.015599170699715614, + "learning_rate": 0.0006, + "loss": 4.0792436599731445, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.015472427010536194, + "learning_rate": 0.0006, + "loss": 4.186446666717529, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.015520439483225346, + "learning_rate": 0.0006, + "loss": 4.246635437011719, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.016998281702399254, + "learning_rate": 0.0006, + "loss": 4.13435173034668, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.016290361061692238, + "learning_rate": 0.0006, + "loss": 4.107428073883057, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.0165516659617424, + "learning_rate": 0.0006, + "loss": 4.1612138748168945, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.01736091636121273, + "learning_rate": 0.0006, + "loss": 4.1072869300842285, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.01787031814455986, + "learning_rate": 0.0006, + "loss": 4.1076812744140625, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.015437985770404339, + "learning_rate": 0.0006, + "loss": 4.097488880157471, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.016306741163134575, + "learning_rate": 0.0006, + "loss": 4.105017185211182, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.015723245218396187, + "learning_rate": 0.0006, + "loss": 4.205283164978027, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.014626456424593925, + "learning_rate": 0.0006, + "loss": 4.086552619934082, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.014348454773426056, + "learning_rate": 0.0006, + "loss": 4.171329975128174, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.015598480589687824, + "learning_rate": 0.0006, + "loss": 4.140815258026123, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.017126062884926796, + "learning_rate": 0.0006, + "loss": 4.157176971435547, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.01860199309885502, + "learning_rate": 0.0006, + "loss": 4.252880096435547, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.021780220791697502, + "learning_rate": 0.0006, + "loss": 4.249861717224121, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.606466293334961, + "eval_runtime": 56.6217, + "eval_samples_per_second": 43.128, + "eval_steps_per_second": 1.36, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.020892828702926636, + "learning_rate": 0.0006, + "loss": 4.102055549621582, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.020306255668401718, + "learning_rate": 0.0006, + "loss": 4.184308052062988, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.02025097794830799, + "learning_rate": 0.0006, + "loss": 3.9691858291625977, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.020254382863640785, + "learning_rate": 0.0006, + "loss": 4.02957820892334, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.019307196140289307, + "learning_rate": 0.0006, + "loss": 4.033181667327881, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.0198476854711771, + "learning_rate": 0.0006, + "loss": 4.0816874504089355, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.019916515797376633, + "learning_rate": 0.0006, + "loss": 4.164889812469482, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.0189144778996706, + "learning_rate": 0.0006, + "loss": 4.038427352905273, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.01754768192768097, + "learning_rate": 0.0006, + "loss": 4.010427474975586, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.016862990334630013, + "learning_rate": 0.0006, + "loss": 4.035970211029053, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.019616752862930298, + "learning_rate": 0.0006, + "loss": 4.079898834228516, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.020131465047597885, + "learning_rate": 0.0006, + "loss": 4.144684791564941, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.02072521299123764, + "learning_rate": 0.0006, + "loss": 4.013660430908203, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.01924615167081356, + "learning_rate": 0.0006, + "loss": 4.071398735046387, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.02013954520225525, + "learning_rate": 0.0006, + "loss": 4.127094268798828, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.02042008936405182, + "learning_rate": 0.0006, + "loss": 4.154261589050293, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.021998876705765724, + "learning_rate": 0.0006, + "loss": 4.204150199890137, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.022946683689951897, + "learning_rate": 0.0006, + "loss": 4.013632774353027, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.02061830274760723, + "learning_rate": 0.0006, + "loss": 4.171084403991699, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.01883152686059475, + "learning_rate": 0.0006, + "loss": 4.149848461151123, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.019945282489061356, + "learning_rate": 0.0006, + "loss": 4.174656867980957, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.020751051604747772, + "learning_rate": 0.0006, + "loss": 4.080747127532959, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.01879691146314144, + "learning_rate": 0.0006, + "loss": 4.057356357574463, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.019037071615457535, + "learning_rate": 0.0006, + "loss": 4.204607009887695, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.01872078701853752, + "learning_rate": 0.0006, + "loss": 4.050586700439453, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.018092811107635498, + "learning_rate": 0.0006, + "loss": 4.102832794189453, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.017010487616062164, + "learning_rate": 0.0006, + "loss": 4.191788673400879, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.01691143773496151, + "learning_rate": 0.0006, + "loss": 4.160456657409668, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.017046816647052765, + "learning_rate": 0.0006, + "loss": 4.073828220367432, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.01707981713116169, + "learning_rate": 0.0006, + "loss": 4.0980095863342285, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.01630452275276184, + "learning_rate": 0.0006, + "loss": 4.101321220397949, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.01576915942132473, + "learning_rate": 0.0006, + "loss": 4.158249855041504, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.016179712489247322, + "learning_rate": 0.0006, + "loss": 4.077852249145508, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.01609169878065586, + "learning_rate": 0.0006, + "loss": 4.187317848205566, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.017148254439234734, + "learning_rate": 0.0006, + "loss": 4.109071731567383, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.015842624008655548, + "learning_rate": 0.0006, + "loss": 4.145096302032471, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.014906194061040878, + "learning_rate": 0.0006, + "loss": 4.028440475463867, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.01485038734972477, + "learning_rate": 0.0006, + "loss": 4.175201416015625, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.015834150835871696, + "learning_rate": 0.0006, + "loss": 4.076434135437012, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.015411358326673508, + "learning_rate": 0.0006, + "loss": 4.03912878036499, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.014762846753001213, + "learning_rate": 0.0006, + "loss": 4.019827365875244, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.015269497409462929, + "learning_rate": 0.0006, + "loss": 4.116957187652588, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.015171617269515991, + "learning_rate": 0.0006, + "loss": 4.2069926261901855, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.014770902693271637, + "learning_rate": 0.0006, + "loss": 4.092021942138672, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.014957531355321407, + "learning_rate": 0.0006, + "loss": 4.16819953918457, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.014774451963603497, + "learning_rate": 0.0006, + "loss": 4.046073913574219, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.014320445246994495, + "learning_rate": 0.0006, + "loss": 3.977792501449585, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.01413130946457386, + "learning_rate": 0.0006, + "loss": 4.130649566650391, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.014336307533085346, + "learning_rate": 0.0006, + "loss": 4.161357402801514, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.014516931027173996, + "learning_rate": 0.0006, + "loss": 4.14248514175415, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.014377126470208168, + "learning_rate": 0.0006, + "loss": 4.062711715698242, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.014332668855786324, + "learning_rate": 0.0006, + "loss": 4.06423807144165, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.014030571095645428, + "learning_rate": 0.0006, + "loss": 4.178613662719727, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.014376776292920113, + "learning_rate": 0.0006, + "loss": 4.073130130767822, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.013827716931700706, + "learning_rate": 0.0006, + "loss": 4.043869972229004, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.015548762865364552, + "learning_rate": 0.0006, + "loss": 4.134049415588379, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.017041970044374466, + "learning_rate": 0.0006, + "loss": 4.205024242401123, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.017638275399804115, + "learning_rate": 0.0006, + "loss": 4.199924468994141, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.017771627753973007, + "learning_rate": 0.0006, + "loss": 4.15706729888916, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.015933562070131302, + "learning_rate": 0.0006, + "loss": 4.220742225646973, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.01668144389986992, + "learning_rate": 0.0006, + "loss": 4.161426544189453, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.018014680594205856, + "learning_rate": 0.0006, + "loss": 4.183046817779541, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.018808670341968536, + "learning_rate": 0.0006, + "loss": 4.105293273925781, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.018003476783633232, + "learning_rate": 0.0006, + "loss": 4.103171348571777, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.016865914687514305, + "learning_rate": 0.0006, + "loss": 4.125634670257568, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.018371153622865677, + "learning_rate": 0.0006, + "loss": 4.192461013793945, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.019045377150177956, + "learning_rate": 0.0006, + "loss": 4.198105812072754, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.01594168320298195, + "learning_rate": 0.0006, + "loss": 4.169788360595703, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.015645820647478104, + "learning_rate": 0.0006, + "loss": 4.047277450561523, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.016313156113028526, + "learning_rate": 0.0006, + "loss": 4.087392807006836, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.017050625756382942, + "learning_rate": 0.0006, + "loss": 4.0908002853393555, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.017730968073010445, + "learning_rate": 0.0006, + "loss": 4.110957622528076, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.563793182373047, + "eval_runtime": 56.5922, + "eval_samples_per_second": 43.151, + "eval_steps_per_second": 1.361, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.015687910839915276, + "learning_rate": 0.0006, + "loss": 4.143589019775391, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.017451366409659386, + "learning_rate": 0.0006, + "loss": 4.0326666831970215, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.01909538544714451, + "learning_rate": 0.0006, + "loss": 3.9706249237060547, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.01988379657268524, + "learning_rate": 0.0006, + "loss": 3.9531009197235107, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.02035115472972393, + "learning_rate": 0.0006, + "loss": 3.962214469909668, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.01952001079916954, + "learning_rate": 0.0006, + "loss": 4.182262897491455, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.01854325644671917, + "learning_rate": 0.0006, + "loss": 4.127663612365723, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.017757616937160492, + "learning_rate": 0.0006, + "loss": 4.100804805755615, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.016712479293346405, + "learning_rate": 0.0006, + "loss": 4.026284217834473, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.0168139711022377, + "learning_rate": 0.0006, + "loss": 4.183882713317871, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.017060423269867897, + "learning_rate": 0.0006, + "loss": 4.131832599639893, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.01726631447672844, + "learning_rate": 0.0006, + "loss": 3.9156057834625244, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.016348622739315033, + "learning_rate": 0.0006, + "loss": 4.1731367111206055, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.016640059649944305, + "learning_rate": 0.0006, + "loss": 4.10997200012207, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.01601342298090458, + "learning_rate": 0.0006, + "loss": 4.081145286560059, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.01735631190240383, + "learning_rate": 0.0006, + "loss": 4.202108860015869, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.019164685159921646, + "learning_rate": 0.0006, + "loss": 4.104109764099121, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.01881759613752365, + "learning_rate": 0.0006, + "loss": 4.145532131195068, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.017656555399298668, + "learning_rate": 0.0006, + "loss": 4.012782096862793, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.01624823734164238, + "learning_rate": 0.0006, + "loss": 4.001727104187012, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.016973139718174934, + "learning_rate": 0.0006, + "loss": 4.095399856567383, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.015503671951591969, + "learning_rate": 0.0006, + "loss": 4.034431457519531, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.015370250679552555, + "learning_rate": 0.0006, + "loss": 4.075836181640625, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.01564021222293377, + "learning_rate": 0.0006, + "loss": 4.125202178955078, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.016717975959181786, + "learning_rate": 0.0006, + "loss": 4.153017044067383, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.015263617038726807, + "learning_rate": 0.0006, + "loss": 4.021814823150635, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.016497142612934113, + "learning_rate": 0.0006, + "loss": 4.069095611572266, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.01669355481863022, + "learning_rate": 0.0006, + "loss": 4.174587726593018, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.019695773720741272, + "learning_rate": 0.0006, + "loss": 4.0073041915893555, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.023720644414424896, + "learning_rate": 0.0006, + "loss": 4.111948013305664, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.02399350143969059, + "learning_rate": 0.0006, + "loss": 4.081164360046387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.019630271941423416, + "learning_rate": 0.0006, + "loss": 4.065731048583984, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.019679008051753044, + "learning_rate": 0.0006, + "loss": 3.993852376937866, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.020597299560904503, + "learning_rate": 0.0006, + "loss": 4.275555610656738, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.01983889564871788, + "learning_rate": 0.0006, + "loss": 4.254708290100098, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.019150668755173683, + "learning_rate": 0.0006, + "loss": 4.097885608673096, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.018847715109586716, + "learning_rate": 0.0006, + "loss": 4.054081439971924, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.019607288762927055, + "learning_rate": 0.0006, + "loss": 3.9891440868377686, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.018863150849938393, + "learning_rate": 0.0006, + "loss": 4.084877014160156, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.017340239137411118, + "learning_rate": 0.0006, + "loss": 4.113290786743164, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.015628693625330925, + "learning_rate": 0.0006, + "loss": 4.1483001708984375, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.016493283212184906, + "learning_rate": 0.0006, + "loss": 4.1040143966674805, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.016732443124055862, + "learning_rate": 0.0006, + "loss": 4.07205057144165, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.017176156863570213, + "learning_rate": 0.0006, + "loss": 4.1279754638671875, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.015730608254671097, + "learning_rate": 0.0006, + "loss": 4.086658000946045, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.016111090779304504, + "learning_rate": 0.0006, + "loss": 4.064335823059082, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.01769379712641239, + "learning_rate": 0.0006, + "loss": 4.116047382354736, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.01683102548122406, + "learning_rate": 0.0006, + "loss": 4.095986843109131, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.015726450830698013, + "learning_rate": 0.0006, + "loss": 4.10775089263916, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.014760235324501991, + "learning_rate": 0.0006, + "loss": 4.1234893798828125, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.014933438040316105, + "learning_rate": 0.0006, + "loss": 4.079200744628906, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.015723997727036476, + "learning_rate": 0.0006, + "loss": 4.07865047454834, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.014539001509547234, + "learning_rate": 0.0006, + "loss": 4.036242485046387, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.014844979159533978, + "learning_rate": 0.0006, + "loss": 4.1051483154296875, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.015939844772219658, + "learning_rate": 0.0006, + "loss": 4.017253875732422, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.014771555550396442, + "learning_rate": 0.0006, + "loss": 4.0694661140441895, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.015777921304106712, + "learning_rate": 0.0006, + "loss": 3.947587490081787, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.017451291903853416, + "learning_rate": 0.0006, + "loss": 4.093625068664551, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.016829246655106544, + "learning_rate": 0.0006, + "loss": 4.09030818939209, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.016984397545456886, + "learning_rate": 0.0006, + "loss": 4.140626907348633, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.01732603646814823, + "learning_rate": 0.0006, + "loss": 4.128276824951172, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.01744619384407997, + "learning_rate": 0.0006, + "loss": 4.103130340576172, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.016331320628523827, + "learning_rate": 0.0006, + "loss": 4.037951469421387, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.015919430181384087, + "learning_rate": 0.0006, + "loss": 4.121513366699219, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.016153695061802864, + "learning_rate": 0.0006, + "loss": 4.099384307861328, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.01628800667822361, + "learning_rate": 0.0006, + "loss": 4.092325687408447, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.017285684123635292, + "learning_rate": 0.0006, + "loss": 3.9828906059265137, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.016047554090619087, + "learning_rate": 0.0006, + "loss": 4.100419044494629, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.01562986709177494, + "learning_rate": 0.0006, + "loss": 4.065832614898682, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.01677163504064083, + "learning_rate": 0.0006, + "loss": 4.077518463134766, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.01755492016673088, + "learning_rate": 0.0006, + "loss": 4.1076154708862305, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.01929762214422226, + "learning_rate": 0.0006, + "loss": 4.09603214263916, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.582442283630371, + "eval_runtime": 62.8649, + "eval_samples_per_second": 38.845, + "eval_steps_per_second": 1.225, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.019129136577248573, + "learning_rate": 0.0006, + "loss": 4.095005989074707, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.019167408347129822, + "learning_rate": 0.0006, + "loss": 4.054426670074463, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.01789460889995098, + "learning_rate": 0.0006, + "loss": 4.0434417724609375, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.015947027131915092, + "learning_rate": 0.0006, + "loss": 4.074516296386719, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.016667252406477928, + "learning_rate": 0.0006, + "loss": 4.108545303344727, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.016014399006962776, + "learning_rate": 0.0006, + "loss": 4.026925086975098, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.015397695824503899, + "learning_rate": 0.0006, + "loss": 4.163970470428467, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.01563289947807789, + "learning_rate": 0.0006, + "loss": 4.116578578948975, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.015512318350374699, + "learning_rate": 0.0006, + "loss": 4.096311569213867, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.015426642261445522, + "learning_rate": 0.0006, + "loss": 4.017065048217773, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.015115435235202312, + "learning_rate": 0.0006, + "loss": 4.005434989929199, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.015235894359648228, + "learning_rate": 0.0006, + "loss": 3.982743263244629, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.016095101833343506, + "learning_rate": 0.0006, + "loss": 4.0849385261535645, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.015201197937130928, + "learning_rate": 0.0006, + "loss": 4.103863716125488, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.015687478706240654, + "learning_rate": 0.0006, + "loss": 4.12074089050293, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.015201589092612267, + "learning_rate": 0.0006, + "loss": 4.056872367858887, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.015154894441366196, + "learning_rate": 0.0006, + "loss": 4.089670181274414, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.01586608588695526, + "learning_rate": 0.0006, + "loss": 4.155317306518555, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.01609654724597931, + "learning_rate": 0.0006, + "loss": 3.995800495147705, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.015495185740292072, + "learning_rate": 0.0006, + "loss": 3.995769739151001, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.017042381688952446, + "learning_rate": 0.0006, + "loss": 4.132832050323486, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.018706614151597023, + "learning_rate": 0.0006, + "loss": 4.038820743560791, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.018316026777029037, + "learning_rate": 0.0006, + "loss": 3.972052574157715, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.018375389277935028, + "learning_rate": 0.0006, + "loss": 4.152097702026367, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.016879886388778687, + "learning_rate": 0.0006, + "loss": 4.020869731903076, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.015728183090686798, + "learning_rate": 0.0006, + "loss": 4.008016586303711, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.016147758811712265, + "learning_rate": 0.0006, + "loss": 4.035854339599609, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.017774797976017, + "learning_rate": 0.0006, + "loss": 4.122506141662598, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.016590069979429245, + "learning_rate": 0.0006, + "loss": 4.082847595214844, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.016544528305530548, + "learning_rate": 0.0006, + "loss": 4.184270858764648, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.017972925677895546, + "learning_rate": 0.0006, + "loss": 3.987558603286743, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.018562227487564087, + "learning_rate": 0.0006, + "loss": 4.1193461418151855, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.018896685913205147, + "learning_rate": 0.0006, + "loss": 4.057808876037598, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.016793809831142426, + "learning_rate": 0.0006, + "loss": 4.069523334503174, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.01698676310479641, + "learning_rate": 0.0006, + "loss": 3.997624158859253, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.016784202307462692, + "learning_rate": 0.0006, + "loss": 4.174795627593994, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.016785014420747757, + "learning_rate": 0.0006, + "loss": 3.999321222305298, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.018615184351801872, + "learning_rate": 0.0006, + "loss": 4.191237926483154, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.019090790301561356, + "learning_rate": 0.0006, + "loss": 4.106966018676758, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.018298320472240448, + "learning_rate": 0.0006, + "loss": 4.063363552093506, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.018541090190410614, + "learning_rate": 0.0006, + "loss": 4.101193428039551, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.01849970407783985, + "learning_rate": 0.0006, + "loss": 4.071335792541504, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.018309077247977257, + "learning_rate": 0.0006, + "loss": 4.245305061340332, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.017681840807199478, + "learning_rate": 0.0006, + "loss": 4.081173419952393, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.016286224126815796, + "learning_rate": 0.0006, + "loss": 4.024351596832275, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.015837235376238823, + "learning_rate": 0.0006, + "loss": 4.028205394744873, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.016062596812844276, + "learning_rate": 0.0006, + "loss": 4.131122589111328, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.01711239665746689, + "learning_rate": 0.0006, + "loss": 4.124416351318359, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.015304913744330406, + "learning_rate": 0.0006, + "loss": 4.056122779846191, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.015971887856721878, + "learning_rate": 0.0006, + "loss": 4.132555961608887, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.015787221491336823, + "learning_rate": 0.0006, + "loss": 4.051596164703369, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.01605444774031639, + "learning_rate": 0.0006, + "loss": 4.191808700561523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.016383009031414986, + "learning_rate": 0.0006, + "loss": 4.0648417472839355, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.016135554760694504, + "learning_rate": 0.0006, + "loss": 3.988563060760498, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.016634158790111542, + "learning_rate": 0.0006, + "loss": 4.133493423461914, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.01688210666179657, + "learning_rate": 0.0006, + "loss": 4.091277122497559, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.016102122142910957, + "learning_rate": 0.0006, + "loss": 4.051219940185547, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.016585038974881172, + "learning_rate": 0.0006, + "loss": 4.058587074279785, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.017700951546430588, + "learning_rate": 0.0006, + "loss": 4.016030311584473, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.017922203987836838, + "learning_rate": 0.0006, + "loss": 4.079977989196777, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.017299890518188477, + "learning_rate": 0.0006, + "loss": 4.062984943389893, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.018648581579327583, + "learning_rate": 0.0006, + "loss": 4.065717697143555, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.019057920202612877, + "learning_rate": 0.0006, + "loss": 4.227892875671387, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.018824422731995583, + "learning_rate": 0.0006, + "loss": 4.116058349609375, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.019084978848695755, + "learning_rate": 0.0006, + "loss": 4.061677932739258, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.017842981964349747, + "learning_rate": 0.0006, + "loss": 4.083813667297363, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.01880858838558197, + "learning_rate": 0.0006, + "loss": 4.152561187744141, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.017400087788701057, + "learning_rate": 0.0006, + "loss": 4.154635429382324, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.017682507634162903, + "learning_rate": 0.0006, + "loss": 4.151128768920898, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.015803052112460136, + "learning_rate": 0.0006, + "loss": 4.10498571395874, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.01582511141896248, + "learning_rate": 0.0006, + "loss": 4.140159606933594, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.01925666816532612, + "learning_rate": 0.0006, + "loss": 4.126394271850586, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.615617275238037, + "eval_runtime": 56.8895, + "eval_samples_per_second": 42.925, + "eval_steps_per_second": 1.354, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.01677853614091873, + "learning_rate": 0.0006, + "loss": 4.102395057678223, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.017901137471199036, + "learning_rate": 0.0006, + "loss": 3.976609230041504, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.017978668212890625, + "learning_rate": 0.0006, + "loss": 3.944416046142578, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.019967492669820786, + "learning_rate": 0.0006, + "loss": 3.992234468460083, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.018937544897198677, + "learning_rate": 0.0006, + "loss": 4.028573036193848, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.016852952539920807, + "learning_rate": 0.0006, + "loss": 4.062442779541016, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.016643647104501724, + "learning_rate": 0.0006, + "loss": 4.08871603012085, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.016872866079211235, + "learning_rate": 0.0006, + "loss": 3.9514245986938477, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.018045669421553612, + "learning_rate": 0.0006, + "loss": 4.038842678070068, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.017853403463959694, + "learning_rate": 0.0006, + "loss": 3.93709135055542, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.017507443204522133, + "learning_rate": 0.0006, + "loss": 3.98665714263916, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.01766853965818882, + "learning_rate": 0.0006, + "loss": 3.937126636505127, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.016660021618008614, + "learning_rate": 0.0006, + "loss": 3.8397014141082764, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.017719309777021408, + "learning_rate": 0.0006, + "loss": 4.003579616546631, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.018234414979815483, + "learning_rate": 0.0006, + "loss": 4.053859710693359, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.018777024000883102, + "learning_rate": 0.0006, + "loss": 4.056182861328125, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.017443502321839333, + "learning_rate": 0.0006, + "loss": 4.025068283081055, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.016500331461429596, + "learning_rate": 0.0006, + "loss": 4.101631164550781, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.016101544722914696, + "learning_rate": 0.0006, + "loss": 4.027082443237305, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.015471171587705612, + "learning_rate": 0.0006, + "loss": 4.022909641265869, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.015779748558998108, + "learning_rate": 0.0006, + "loss": 4.18796968460083, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.015522679314017296, + "learning_rate": 0.0006, + "loss": 4.005777359008789, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.015226504765450954, + "learning_rate": 0.0006, + "loss": 4.108416557312012, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.015095778740942478, + "learning_rate": 0.0006, + "loss": 4.08029842376709, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.015601031482219696, + "learning_rate": 0.0006, + "loss": 3.989492416381836, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.014825052581727505, + "learning_rate": 0.0006, + "loss": 3.998520612716675, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.014821100980043411, + "learning_rate": 0.0006, + "loss": 4.001777172088623, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.015333754941821098, + "learning_rate": 0.0006, + "loss": 4.11985969543457, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.015940699726343155, + "learning_rate": 0.0006, + "loss": 4.025373935699463, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.015096287243068218, + "learning_rate": 0.0006, + "loss": 4.068551063537598, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.01533548254519701, + "learning_rate": 0.0006, + "loss": 3.9800801277160645, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.014648732729256153, + "learning_rate": 0.0006, + "loss": 3.9670190811157227, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.015759671106934547, + "learning_rate": 0.0006, + "loss": 4.050443649291992, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.016054073348641396, + "learning_rate": 0.0006, + "loss": 4.22734260559082, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.0160264540463686, + "learning_rate": 0.0006, + "loss": 4.116174221038818, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.015426610596477985, + "learning_rate": 0.0006, + "loss": 4.000178337097168, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.01520642638206482, + "learning_rate": 0.0006, + "loss": 4.218410968780518, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.015305949375033379, + "learning_rate": 0.0006, + "loss": 4.025891304016113, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.015461564064025879, + "learning_rate": 0.0006, + "loss": 4.100950717926025, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.01712625101208687, + "learning_rate": 0.0006, + "loss": 4.0620856285095215, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.017133301123976707, + "learning_rate": 0.0006, + "loss": 4.097138404846191, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.01683625392615795, + "learning_rate": 0.0006, + "loss": 4.084364891052246, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.0165710486471653, + "learning_rate": 0.0006, + "loss": 4.042361259460449, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.01660393364727497, + "learning_rate": 0.0006, + "loss": 4.1395487785339355, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.017025845125317574, + "learning_rate": 0.0006, + "loss": 4.1207075119018555, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.01660497486591339, + "learning_rate": 0.0006, + "loss": 4.103846073150635, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.01607946865260601, + "learning_rate": 0.0006, + "loss": 4.189016342163086, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.01614505983889103, + "learning_rate": 0.0006, + "loss": 4.079860687255859, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.017085086554288864, + "learning_rate": 0.0006, + "loss": 3.9774293899536133, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.01816520094871521, + "learning_rate": 0.0006, + "loss": 4.113666534423828, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.016597526147961617, + "learning_rate": 0.0006, + "loss": 4.068242073059082, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.01672235131263733, + "learning_rate": 0.0006, + "loss": 4.024444580078125, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.020093556493520737, + "learning_rate": 0.0006, + "loss": 4.064124584197998, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.023141304031014442, + "learning_rate": 0.0006, + "loss": 4.130439758300781, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.02289372682571411, + "learning_rate": 0.0006, + "loss": 4.14794921875, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.01919020712375641, + "learning_rate": 0.0006, + "loss": 4.123961925506592, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.016932329162955284, + "learning_rate": 0.0006, + "loss": 4.0944132804870605, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.016644535586237907, + "learning_rate": 0.0006, + "loss": 4.124269962310791, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.016302216798067093, + "learning_rate": 0.0006, + "loss": 4.0267205238342285, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.01673300378024578, + "learning_rate": 0.0006, + "loss": 4.10650634765625, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.016581356525421143, + "learning_rate": 0.0006, + "loss": 4.1547956466674805, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.01691797934472561, + "learning_rate": 0.0006, + "loss": 4.056936740875244, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.01596726104617119, + "learning_rate": 0.0006, + "loss": 3.9891414642333984, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.015615091659128666, + "learning_rate": 0.0006, + "loss": 4.137948036193848, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.015703413635492325, + "learning_rate": 0.0006, + "loss": 3.9844930171966553, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.015453094616532326, + "learning_rate": 0.0006, + "loss": 4.098443508148193, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.01449158601462841, + "learning_rate": 0.0006, + "loss": 4.090852737426758, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.01537515688687563, + "learning_rate": 0.0006, + "loss": 4.239612579345703, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.015989921987056732, + "learning_rate": 0.0006, + "loss": 4.142073154449463, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.015985611826181412, + "learning_rate": 0.0006, + "loss": 4.100076675415039, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.015260833315551281, + "learning_rate": 0.0006, + "loss": 4.068576335906982, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.018047941848635674, + "learning_rate": 0.0006, + "loss": 4.005878925323486, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.622848987579346, + "eval_runtime": 56.9618, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.015976838767528534, + "learning_rate": 0.0006, + "loss": 4.035689353942871, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.015176220797002316, + "learning_rate": 0.0006, + "loss": 3.932197332382202, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.016676228493452072, + "learning_rate": 0.0006, + "loss": 4.014171123504639, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.01698281429708004, + "learning_rate": 0.0006, + "loss": 4.117893218994141, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.017872486263513565, + "learning_rate": 0.0006, + "loss": 4.132119178771973, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.017615554854273796, + "learning_rate": 0.0006, + "loss": 4.072000980377197, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.017469394952058792, + "learning_rate": 0.0006, + "loss": 4.017726421356201, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.01954762637615204, + "learning_rate": 0.0006, + "loss": 4.014443874359131, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.018935853615403175, + "learning_rate": 0.0006, + "loss": 4.0042948722839355, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.019382692873477936, + "learning_rate": 0.0006, + "loss": 4.06160831451416, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.019677115604281425, + "learning_rate": 0.0006, + "loss": 4.13466215133667, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.020365122705698013, + "learning_rate": 0.0006, + "loss": 4.095000267028809, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.021842138841748238, + "learning_rate": 0.0006, + "loss": 3.9877853393554688, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.02354881353676319, + "learning_rate": 0.0006, + "loss": 3.9430112838745117, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.026602735742926598, + "learning_rate": 0.0006, + "loss": 4.012076377868652, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.023976484313607216, + "learning_rate": 0.0006, + "loss": 3.9768595695495605, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.02336616814136505, + "learning_rate": 0.0006, + "loss": 4.1096038818359375, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.02426023595035076, + "learning_rate": 0.0006, + "loss": 4.157189846038818, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.021572696045041084, + "learning_rate": 0.0006, + "loss": 3.979175090789795, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.08300432562828064, + "learning_rate": 0.0006, + "loss": 4.092336654663086, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.7323967814445496, + "learning_rate": 0.0006, + "loss": 5.337990760803223, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.3363337814807892, + "learning_rate": 0.0006, + "loss": 6.40043830871582, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.598686933517456, + "learning_rate": 0.0006, + "loss": 7.004765510559082, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.5998364686965942, + "learning_rate": 0.0006, + "loss": 8.184747695922852, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 1.408208966255188, + "learning_rate": 0.0006, + "loss": 8.88978385925293, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.7829785943031311, + "learning_rate": 0.0006, + "loss": 11.13601303100586, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.7769215106964111, + "learning_rate": 0.0006, + "loss": 9.90416145324707, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.46508079767227173, + "learning_rate": 0.0006, + "loss": 9.647209167480469, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 1.0228968858718872, + "learning_rate": 0.0006, + "loss": 11.685434341430664, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.20553778111934662, + "learning_rate": 0.0006, + "loss": 8.621708869934082, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.18466424942016602, + "learning_rate": 0.0006, + "loss": 8.590787887573242, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.10458594560623169, + "learning_rate": 0.0006, + "loss": 8.324458122253418, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.1096842959523201, + "learning_rate": 0.0006, + "loss": 8.300851821899414, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.10501226782798767, + "learning_rate": 0.0006, + "loss": 8.21009635925293, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.13704165816307068, + "learning_rate": 0.0006, + "loss": 8.141355514526367, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.12129108607769012, + "learning_rate": 0.0006, + "loss": 8.074612617492676, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.10028897225856781, + "learning_rate": 0.0006, + "loss": 8.00384521484375, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.11764246970415115, + "learning_rate": 0.0006, + "loss": 8.078598022460938, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.10081959515810013, + "learning_rate": 0.0006, + "loss": 8.008938789367676, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.1003170907497406, + "learning_rate": 0.0006, + "loss": 7.962940216064453, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.08763492107391357, + "learning_rate": 0.0006, + "loss": 7.909419536590576, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0636807233095169, + "learning_rate": 0.0006, + "loss": 7.878905773162842, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.06005936488509178, + "learning_rate": 0.0006, + "loss": 7.850980758666992, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.06415296345949173, + "learning_rate": 0.0006, + "loss": 7.830353260040283, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.0732007846236229, + "learning_rate": 0.0006, + "loss": 7.830746650695801, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.06628110259771347, + "learning_rate": 0.0006, + "loss": 7.82661771774292, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.05544218048453331, + "learning_rate": 0.0006, + "loss": 7.814305782318115, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.050327103585004807, + "learning_rate": 0.0006, + "loss": 7.764055252075195, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.04918425530195236, + "learning_rate": 0.0006, + "loss": 7.744625091552734, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.042862266302108765, + "learning_rate": 0.0006, + "loss": 7.764426231384277, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.04338089004158974, + "learning_rate": 0.0006, + "loss": 7.712530136108398, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.04127737879753113, + "learning_rate": 0.0006, + "loss": 7.742180347442627, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.03358631581068039, + "learning_rate": 0.0006, + "loss": 7.714834213256836, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.02845790982246399, + "learning_rate": 0.0006, + "loss": 7.7021026611328125, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.03012419492006302, + "learning_rate": 0.0006, + "loss": 7.712377548217773, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.032981179654598236, + "learning_rate": 0.0006, + "loss": 7.666385650634766, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.03357074037194252, + "learning_rate": 0.0006, + "loss": 7.649097919464111, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.030914708971977234, + "learning_rate": 0.0006, + "loss": 7.649196624755859, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.026296131312847137, + "learning_rate": 0.0006, + "loss": 7.629964828491211, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.026677245274186134, + "learning_rate": 0.0006, + "loss": 7.599973678588867, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.023208219558000565, + "learning_rate": 0.0006, + "loss": 7.601624488830566, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.020741475746035576, + "learning_rate": 0.0006, + "loss": 7.603902339935303, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.019221751019358635, + "learning_rate": 0.0006, + "loss": 7.5535888671875, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.019491951912641525, + "learning_rate": 0.0006, + "loss": 7.54597282409668, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.019702572375535965, + "learning_rate": 0.0006, + "loss": 7.537567138671875, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.01768800988793373, + "learning_rate": 0.0006, + "loss": 7.512800216674805, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.01926138624548912, + "learning_rate": 0.0006, + "loss": 7.509932518005371, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.019030006602406502, + "learning_rate": 0.0006, + "loss": 7.497983932495117, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.017313841730356216, + "learning_rate": 0.0006, + "loss": 7.463518142700195, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.01310762483626604, + "learning_rate": 0.0006, + "loss": 7.474668502807617, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.013008282519876957, + "learning_rate": 0.0006, + "loss": 7.45982551574707, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.01441363524645567, + "learning_rate": 0.0006, + "loss": 7.446529388427734, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 7.486120223999023, + "eval_runtime": 70.4986, + "eval_samples_per_second": 34.639, + "eval_steps_per_second": 1.092, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.014061855152249336, + "learning_rate": 0.0006, + "loss": 7.4692182540893555, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.01587500050663948, + "learning_rate": 0.0006, + "loss": 7.4182024002075195, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.014783984050154686, + "learning_rate": 0.0006, + "loss": 7.466778755187988, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.011560854502022266, + "learning_rate": 0.0006, + "loss": 7.438002586364746, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.00970957800745964, + "learning_rate": 0.0006, + "loss": 7.442694187164307, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.009633459150791168, + "learning_rate": 0.0006, + "loss": 7.458884239196777, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.010664350353181362, + "learning_rate": 0.0006, + "loss": 7.48378324508667, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.011440591886639595, + "learning_rate": 0.0006, + "loss": 7.474484920501709, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.01150908600538969, + "learning_rate": 0.0006, + "loss": 7.417902946472168, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.010375697165727615, + "learning_rate": 0.0006, + "loss": 7.403279781341553, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.009075703099370003, + "learning_rate": 0.0006, + "loss": 7.447347164154053, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.008892177604138851, + "learning_rate": 0.0006, + "loss": 7.402681350708008, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.009492084383964539, + "learning_rate": 0.0006, + "loss": 7.3871355056762695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.009139460511505604, + "learning_rate": 0.0006, + "loss": 7.419617176055908, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.007959138602018356, + "learning_rate": 0.0006, + "loss": 7.374949932098389, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.007571548223495483, + "learning_rate": 0.0006, + "loss": 7.4205241203308105, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.007885400205850601, + "learning_rate": 0.0006, + "loss": 7.416543006896973, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.008650158531963825, + "learning_rate": 0.0006, + "loss": 7.357326984405518, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007483262103050947, + "learning_rate": 0.0006, + "loss": 7.410891056060791, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.00869277399033308, + "learning_rate": 0.0006, + "loss": 7.392127990722656, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.00885078776627779, + "learning_rate": 0.0006, + "loss": 7.405449867248535, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.00958226341754198, + "learning_rate": 0.0006, + "loss": 7.399868488311768, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.006991599686443806, + "learning_rate": 0.0006, + "loss": 7.35369873046875, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.007415142375975847, + "learning_rate": 0.0006, + "loss": 7.358353137969971, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.007024371065199375, + "learning_rate": 0.0006, + "loss": 7.4070892333984375, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.007234332151710987, + "learning_rate": 0.0006, + "loss": 7.379120349884033, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.00798892229795456, + "learning_rate": 0.0006, + "loss": 7.365455627441406, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.007303398102521896, + "learning_rate": 0.0006, + "loss": 7.341798782348633, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.007119669578969479, + "learning_rate": 0.0006, + "loss": 7.363280296325684, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.007457619532942772, + "learning_rate": 0.0006, + "loss": 7.39388370513916, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007905688136816025, + "learning_rate": 0.0006, + "loss": 7.380652904510498, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006212855689227581, + "learning_rate": 0.0006, + "loss": 7.398769378662109, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.0070641543716192245, + "learning_rate": 0.0006, + "loss": 7.3503546714782715, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.00787805300205946, + "learning_rate": 0.0006, + "loss": 7.347678184509277, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006760426796972752, + "learning_rate": 0.0006, + "loss": 7.380006790161133, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.00772385997697711, + "learning_rate": 0.0006, + "loss": 7.363680362701416, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.007373564876616001, + "learning_rate": 0.0006, + "loss": 7.333014488220215, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006717653013765812, + "learning_rate": 0.0006, + "loss": 7.360076427459717, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.007191413082182407, + "learning_rate": 0.0006, + "loss": 7.397030830383301, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.007831960916519165, + "learning_rate": 0.0006, + "loss": 7.369959831237793, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006947481073439121, + "learning_rate": 0.0006, + "loss": 7.333254337310791, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.006243606563657522, + "learning_rate": 0.0006, + "loss": 7.327378273010254, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006298339460045099, + "learning_rate": 0.0006, + "loss": 7.323482036590576, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.007587412837892771, + "learning_rate": 0.0006, + "loss": 7.313789367675781, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.007140746805816889, + "learning_rate": 0.0006, + "loss": 7.356067657470703, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.008315625600516796, + "learning_rate": 0.0006, + "loss": 7.3339643478393555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.006547102238982916, + "learning_rate": 0.0006, + "loss": 7.342191696166992, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006068122107535601, + "learning_rate": 0.0006, + "loss": 7.321015357971191, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.007316991221159697, + "learning_rate": 0.0006, + "loss": 7.346490859985352, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.0067388564348220825, + "learning_rate": 0.0006, + "loss": 7.340723991394043, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.006839963141828775, + "learning_rate": 0.0006, + "loss": 7.329365253448486, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.00805041566491127, + "learning_rate": 0.0006, + "loss": 7.298934459686279, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.007197503466159105, + "learning_rate": 0.0006, + "loss": 7.331496238708496, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.007220805156975985, + "learning_rate": 0.0006, + "loss": 7.317695140838623, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.00874276738613844, + "learning_rate": 0.0006, + "loss": 7.272614479064941, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.008386565372347832, + "learning_rate": 0.0006, + "loss": 7.295244216918945, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.0077730282209813595, + "learning_rate": 0.0006, + "loss": 7.3395490646362305, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.008441480807960033, + "learning_rate": 0.0006, + "loss": 7.319921493530273, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.007398269604891539, + "learning_rate": 0.0006, + "loss": 7.3029465675354, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.006878677289932966, + "learning_rate": 0.0006, + "loss": 7.318467140197754, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.007616598624736071, + "learning_rate": 0.0006, + "loss": 7.3112077713012695, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.011629699729382992, + "learning_rate": 0.0006, + "loss": 7.32523250579834, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.011011249385774136, + "learning_rate": 0.0006, + "loss": 7.287247657775879, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.010208295658230782, + "learning_rate": 0.0006, + "loss": 7.292117118835449, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.008651095442473888, + "learning_rate": 0.0006, + "loss": 7.302201271057129, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.006692920345813036, + "learning_rate": 0.0006, + "loss": 7.319568634033203, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.010033348575234413, + "learning_rate": 0.0006, + "loss": 7.2738566398620605, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.011790832504630089, + "learning_rate": 0.0006, + "loss": 7.298540115356445, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.018622662872076035, + "learning_rate": 0.0006, + "loss": 7.267182350158691, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.02520257607102394, + "learning_rate": 0.0006, + "loss": 7.261566162109375, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.036332644522190094, + "learning_rate": 0.0006, + "loss": 7.270719051361084, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.04717305675148964, + "learning_rate": 0.0006, + "loss": 7.293332099914551, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 7.354280471801758, + "eval_runtime": 56.4353, + "eval_samples_per_second": 43.271, + "eval_steps_per_second": 1.364, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.06661626696586609, + "learning_rate": 0.0006, + "loss": 7.338189601898193, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.05452054366469383, + "learning_rate": 0.0006, + "loss": 7.316764831542969, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.025638775900006294, + "learning_rate": 0.0006, + "loss": 7.34204626083374, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.022398799657821655, + "learning_rate": 0.0006, + "loss": 7.262149333953857, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.03713499382138252, + "learning_rate": 0.0006, + "loss": 7.27695369720459, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.014949247241020203, + "learning_rate": 0.0006, + "loss": 7.296508312225342, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.02703995071351528, + "learning_rate": 0.0006, + "loss": 7.258018493652344, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.020352816209197044, + "learning_rate": 0.0006, + "loss": 7.241450309753418, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.014578479342162609, + "learning_rate": 0.0006, + "loss": 7.257323265075684, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.026053044945001602, + "learning_rate": 0.0006, + "loss": 7.254680633544922, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.015120275318622589, + "learning_rate": 0.0006, + "loss": 7.261074066162109, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.016822177916765213, + "learning_rate": 0.0006, + "loss": 7.279064178466797, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.016367588192224503, + "learning_rate": 0.0006, + "loss": 7.251429557800293, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.014350418001413345, + "learning_rate": 0.0006, + "loss": 7.259264945983887, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.017078228294849396, + "learning_rate": 0.0006, + "loss": 7.227105140686035, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.011759317480027676, + "learning_rate": 0.0006, + "loss": 7.275888442993164, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.013019340112805367, + "learning_rate": 0.0006, + "loss": 7.2517008781433105, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.01178441196680069, + "learning_rate": 0.0006, + "loss": 7.23576545715332, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.014518891461193562, + "learning_rate": 0.0006, + "loss": 7.253895282745361, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.018159950152039528, + "learning_rate": 0.0006, + "loss": 7.278058052062988, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.007587796077132225, + "learning_rate": 0.0006, + "loss": 7.278578758239746, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.015994979068636894, + "learning_rate": 0.0006, + "loss": 7.239841461181641, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.017830168828368187, + "learning_rate": 0.0006, + "loss": 7.2443976402282715, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.008738759905099869, + "learning_rate": 0.0006, + "loss": 7.232264518737793, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.0175449438393116, + "learning_rate": 0.0006, + "loss": 7.228000640869141, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.019611820578575134, + "learning_rate": 0.0006, + "loss": 7.234999179840088, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.015302863903343678, + "learning_rate": 0.0006, + "loss": 7.242547988891602, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.025919731706380844, + "learning_rate": 0.0006, + "loss": 7.216652870178223, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.034587468951940536, + "learning_rate": 0.0006, + "loss": 7.245336532592773, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.019198840484023094, + "learning_rate": 0.0006, + "loss": 7.256758689880371, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.012316389009356499, + "learning_rate": 0.0006, + "loss": 7.209102630615234, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.023512162268161774, + "learning_rate": 0.0006, + "loss": 7.250742435455322, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.025457851588726044, + "learning_rate": 0.0006, + "loss": 7.263757705688477, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.02192617766559124, + "learning_rate": 0.0006, + "loss": 7.246977806091309, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.02541281282901764, + "learning_rate": 0.0006, + "loss": 7.232361316680908, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.036353763192892075, + "learning_rate": 0.0006, + "loss": 7.242556571960449, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.04555872827768326, + "learning_rate": 0.0006, + "loss": 7.228940963745117, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.04563378915190697, + "learning_rate": 0.0006, + "loss": 7.235276222229004, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.04275760427117348, + "learning_rate": 0.0006, + "loss": 7.238895416259766, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.04535885155200958, + "learning_rate": 0.0006, + "loss": 7.205336570739746, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.015534617006778717, + "learning_rate": 0.0006, + "loss": 7.238393783569336, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.03286299481987953, + "learning_rate": 0.0006, + "loss": 7.216453552246094, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.025453168898820877, + "learning_rate": 0.0006, + "loss": 7.203750133514404, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.021501649171113968, + "learning_rate": 0.0006, + "loss": 7.241621971130371, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.02743634767830372, + "learning_rate": 0.0006, + "loss": 7.22076940536499, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.020890483632683754, + "learning_rate": 0.0006, + "loss": 7.218412399291992, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.022499850019812584, + "learning_rate": 0.0006, + "loss": 7.217696189880371, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.01653696969151497, + "learning_rate": 0.0006, + "loss": 7.156417369842529, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.01834225095808506, + "learning_rate": 0.0006, + "loss": 7.211247444152832, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.020137229934334755, + "learning_rate": 0.0006, + "loss": 7.232499599456787, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.012671459466218948, + "learning_rate": 0.0006, + "loss": 7.172441482543945, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.016803130507469177, + "learning_rate": 0.0006, + "loss": 7.152103424072266, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.01623712107539177, + "learning_rate": 0.0006, + "loss": 7.187032699584961, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.015938356518745422, + "learning_rate": 0.0006, + "loss": 7.171140670776367, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.02580736204981804, + "learning_rate": 0.0006, + "loss": 7.204472541809082, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.023372117429971695, + "learning_rate": 0.0006, + "loss": 7.165554523468018, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.04440586268901825, + "learning_rate": 0.0006, + "loss": 7.169431209564209, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.06797406077384949, + "learning_rate": 0.0006, + "loss": 7.168193340301514, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.0914321169257164, + "learning_rate": 0.0006, + "loss": 7.204671859741211, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.0506940595805645, + "learning_rate": 0.0006, + "loss": 7.15949821472168, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.04819890484213829, + "learning_rate": 0.0006, + "loss": 7.192837238311768, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.040044233202934265, + "learning_rate": 0.0006, + "loss": 7.1600165367126465, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.02573299966752529, + "learning_rate": 0.0006, + "loss": 7.153427600860596, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.046942535787820816, + "learning_rate": 0.0006, + "loss": 7.16767692565918, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.02683161571621895, + "learning_rate": 0.0006, + "loss": 7.17144250869751, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.05363674834370613, + "learning_rate": 0.0006, + "loss": 7.120128631591797, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.023756247013807297, + "learning_rate": 0.0006, + "loss": 7.169703006744385, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.04216783866286278, + "learning_rate": 0.0006, + "loss": 7.1206865310668945, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.02219565585255623, + "learning_rate": 0.0006, + "loss": 7.157215118408203, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.03527127206325531, + "learning_rate": 0.0006, + "loss": 7.166672706604004, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.017748817801475525, + "learning_rate": 0.0006, + "loss": 7.154747009277344, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.033165838569402695, + "learning_rate": 0.0006, + "loss": 7.150866508483887, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 7.144403457641602, + "eval_runtime": 56.8853, + "eval_samples_per_second": 42.928, + "eval_steps_per_second": 1.354, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.015199829824268818, + "learning_rate": 0.0006, + "loss": 7.122735023498535, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.03220256045460701, + "learning_rate": 0.0006, + "loss": 7.143553733825684, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.014687920920550823, + "learning_rate": 0.0006, + "loss": 7.080960750579834, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.023522932082414627, + "learning_rate": 0.0006, + "loss": 7.137356758117676, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.014808216132223606, + "learning_rate": 0.0006, + "loss": 7.09143590927124, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.029339246451854706, + "learning_rate": 0.0006, + "loss": 7.122198104858398, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.022953078150749207, + "learning_rate": 0.0006, + "loss": 7.092505931854248, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.04590052738785744, + "learning_rate": 0.0006, + "loss": 7.113100051879883, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.0524442233145237, + "learning_rate": 0.0006, + "loss": 7.147022247314453, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.0315794013440609, + "learning_rate": 0.0006, + "loss": 7.094045162200928, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.05112722888588905, + "learning_rate": 0.0006, + "loss": 7.054388523101807, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.03146776929497719, + "learning_rate": 0.0006, + "loss": 7.116674423217773, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.03391246497631073, + "learning_rate": 0.0006, + "loss": 7.150345802307129, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.026957903057336807, + "learning_rate": 0.0006, + "loss": 7.058531284332275, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.041532568633556366, + "learning_rate": 0.0006, + "loss": 7.085498809814453, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.020001588389277458, + "learning_rate": 0.0006, + "loss": 7.131487846374512, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.02268712781369686, + "learning_rate": 0.0006, + "loss": 7.0790863037109375, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.023363925516605377, + "learning_rate": 0.0006, + "loss": 7.067816734313965, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.019367508590221405, + "learning_rate": 0.0006, + "loss": 7.042041778564453, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.02745090238749981, + "learning_rate": 0.0006, + "loss": 7.064733028411865, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.02543802745640278, + "learning_rate": 0.0006, + "loss": 7.058948040008545, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.018874092027544975, + "learning_rate": 0.0006, + "loss": 7.009946346282959, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.024662822484970093, + "learning_rate": 0.0006, + "loss": 7.015913963317871, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.02441416308283806, + "learning_rate": 0.0006, + "loss": 7.061622142791748, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.031887322664260864, + "learning_rate": 0.0006, + "loss": 7.046883583068848, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.021930739283561707, + "learning_rate": 0.0006, + "loss": 7.0309624671936035, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.02722783200442791, + "learning_rate": 0.0006, + "loss": 7.072924613952637, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.01483179721981287, + "learning_rate": 0.0006, + "loss": 7.028624534606934, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.024793945252895355, + "learning_rate": 0.0006, + "loss": 7.001623153686523, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.020148713141679764, + "learning_rate": 0.0006, + "loss": 6.9938812255859375, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.02413850650191307, + "learning_rate": 0.0006, + "loss": 7.001243591308594, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.028549203649163246, + "learning_rate": 0.0006, + "loss": 6.987671375274658, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.025988785549998283, + "learning_rate": 0.0006, + "loss": 6.9551167488098145, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.027140729129314423, + "learning_rate": 0.0006, + "loss": 6.9547810554504395, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.030130082741379738, + "learning_rate": 0.0006, + "loss": 6.993186950683594, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.025996994227170944, + "learning_rate": 0.0006, + "loss": 6.984624862670898, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.02423008903861046, + "learning_rate": 0.0006, + "loss": 6.98813009262085, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.029665837064385414, + "learning_rate": 0.0006, + "loss": 6.986610412597656, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.028768595308065414, + "learning_rate": 0.0006, + "loss": 6.898797035217285, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.03111208975315094, + "learning_rate": 0.0006, + "loss": 6.923741817474365, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.0368187390267849, + "learning_rate": 0.0006, + "loss": 6.994343280792236, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.04076763987541199, + "learning_rate": 0.0006, + "loss": 6.909458160400391, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.03281061723828316, + "learning_rate": 0.0006, + "loss": 6.924851417541504, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.025486988946795464, + "learning_rate": 0.0006, + "loss": 6.962926864624023, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.04583485797047615, + "learning_rate": 0.0006, + "loss": 6.876555442810059, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.05460785701870918, + "learning_rate": 0.0006, + "loss": 6.952932357788086, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.040917906910181046, + "learning_rate": 0.0006, + "loss": 6.9624924659729, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.04009668901562691, + "learning_rate": 0.0006, + "loss": 6.94033145904541, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.04500625655055046, + "learning_rate": 0.0006, + "loss": 6.961839199066162, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.03182258829474449, + "learning_rate": 0.0006, + "loss": 6.930580139160156, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.053621504455804825, + "learning_rate": 0.0006, + "loss": 6.950235366821289, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.031919289380311966, + "learning_rate": 0.0006, + "loss": 6.895111560821533, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.042713891714811325, + "learning_rate": 0.0006, + "loss": 6.940885543823242, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.02385653182864189, + "learning_rate": 0.0006, + "loss": 6.948801517486572, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.04100663587450981, + "learning_rate": 0.0006, + "loss": 6.970411777496338, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.02729583904147148, + "learning_rate": 0.0006, + "loss": 6.899379253387451, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.03128620237112045, + "learning_rate": 0.0006, + "loss": 6.931319236755371, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.025488659739494324, + "learning_rate": 0.0006, + "loss": 6.8486409187316895, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.03679376840591431, + "learning_rate": 0.0006, + "loss": 6.840811252593994, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.020578034222126007, + "learning_rate": 0.0006, + "loss": 6.9088616371154785, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.02935841865837574, + "learning_rate": 0.0006, + "loss": 6.806676864624023, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.023319244384765625, + "learning_rate": 0.0006, + "loss": 6.878457546234131, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.02779589593410492, + "learning_rate": 0.0006, + "loss": 6.822447776794434, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.01879825070500374, + "learning_rate": 0.0006, + "loss": 6.854121208190918, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.023286139592528343, + "learning_rate": 0.0006, + "loss": 6.836827278137207, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.030358297750353813, + "learning_rate": 0.0006, + "loss": 6.860037803649902, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.029621105641126633, + "learning_rate": 0.0006, + "loss": 6.884025573730469, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.029829824343323708, + "learning_rate": 0.0006, + "loss": 6.8702392578125, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.036892473697662354, + "learning_rate": 0.0006, + "loss": 6.806652069091797, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.03206709027290344, + "learning_rate": 0.0006, + "loss": 6.828495979309082, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.02744000405073166, + "learning_rate": 0.0006, + "loss": 6.832002639770508, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.02205660194158554, + "learning_rate": 0.0006, + "loss": 6.898828506469727, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 6.8150553703308105, + "eval_runtime": 56.784, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.030421260744333267, + "learning_rate": 0.0006, + "loss": 6.775057792663574, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.0465611033141613, + "learning_rate": 0.0006, + "loss": 6.849390506744385, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.05977972596883774, + "learning_rate": 0.0006, + "loss": 6.725576400756836, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.05432697385549545, + "learning_rate": 0.0006, + "loss": 6.768024444580078, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.04850967228412628, + "learning_rate": 0.0006, + "loss": 6.799457550048828, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.06201786920428276, + "learning_rate": 0.0006, + "loss": 6.897234916687012, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.04695591703057289, + "learning_rate": 0.0006, + "loss": 6.826470375061035, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.03850164636969566, + "learning_rate": 0.0006, + "loss": 6.835901260375977, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.04250155761837959, + "learning_rate": 0.0006, + "loss": 6.782310485839844, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.041465479880571365, + "learning_rate": 0.0006, + "loss": 6.785732746124268, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.02527807280421257, + "learning_rate": 0.0006, + "loss": 6.78061580657959, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.03660374879837036, + "learning_rate": 0.0006, + "loss": 6.783992767333984, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.027567964047193527, + "learning_rate": 0.0006, + "loss": 6.765413284301758, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.024301424622535706, + "learning_rate": 0.0006, + "loss": 6.760368347167969, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.0266821701079607, + "learning_rate": 0.0006, + "loss": 6.71417236328125, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.028125090524554253, + "learning_rate": 0.0006, + "loss": 6.646126747131348, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.0185737032443285, + "learning_rate": 0.0006, + "loss": 6.712944030761719, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.025836674496531487, + "learning_rate": 0.0006, + "loss": 6.707030296325684, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.029744870960712433, + "learning_rate": 0.0006, + "loss": 6.6271071434021, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.03021426685154438, + "learning_rate": 0.0006, + "loss": 6.677826404571533, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.032727014273405075, + "learning_rate": 0.0006, + "loss": 6.657764434814453, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.03696778789162636, + "learning_rate": 0.0006, + "loss": 6.632883548736572, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.034630827605724335, + "learning_rate": 0.0006, + "loss": 6.759583473205566, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.030517758801579475, + "learning_rate": 0.0006, + "loss": 6.702268600463867, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.04105190932750702, + "learning_rate": 0.0006, + "loss": 6.705792427062988, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.06659174710512161, + "learning_rate": 0.0006, + "loss": 6.73862361907959, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.08013619482517242, + "learning_rate": 0.0006, + "loss": 6.710282802581787, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.049850963056087494, + "learning_rate": 0.0006, + "loss": 6.684974670410156, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.061247408390045166, + "learning_rate": 0.0006, + "loss": 6.716034412384033, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.05868857726454735, + "learning_rate": 0.0006, + "loss": 6.589107513427734, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.048878561705350876, + "learning_rate": 0.0006, + "loss": 6.694855690002441, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.028748324140906334, + "learning_rate": 0.0006, + "loss": 6.629969596862793, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.04060306400060654, + "learning_rate": 0.0006, + "loss": 6.721431255340576, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.03833824768662453, + "learning_rate": 0.0006, + "loss": 6.625029563903809, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.03791997581720352, + "learning_rate": 0.0006, + "loss": 6.669187545776367, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.03259684890508652, + "learning_rate": 0.0006, + "loss": 6.665643692016602, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.040658626705408096, + "learning_rate": 0.0006, + "loss": 6.639796257019043, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.03691193461418152, + "learning_rate": 0.0006, + "loss": 6.692244529724121, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.02814812771975994, + "learning_rate": 0.0006, + "loss": 6.61978006362915, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.022155074402689934, + "learning_rate": 0.0006, + "loss": 6.625905990600586, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.030794885009527206, + "learning_rate": 0.0006, + "loss": 6.678500652313232, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.03201768547296524, + "learning_rate": 0.0006, + "loss": 6.6015729904174805, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.025138361379504204, + "learning_rate": 0.0006, + "loss": 6.568234443664551, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.02969353087246418, + "learning_rate": 0.0006, + "loss": 6.588366508483887, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.03136962279677391, + "learning_rate": 0.0006, + "loss": 6.479961395263672, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.030178170651197433, + "learning_rate": 0.0006, + "loss": 6.622910976409912, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.02682245895266533, + "learning_rate": 0.0006, + "loss": 6.597707748413086, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.025154557079076767, + "learning_rate": 0.0006, + "loss": 6.630541801452637, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.03564463183283806, + "learning_rate": 0.0006, + "loss": 6.542823791503906, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.049477651715278625, + "learning_rate": 0.0006, + "loss": 6.646030426025391, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.06536535918712616, + "learning_rate": 0.0006, + "loss": 6.595463275909424, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.07710316777229309, + "learning_rate": 0.0006, + "loss": 6.55524206161499, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.09571163356304169, + "learning_rate": 0.0006, + "loss": 6.659764289855957, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.055076830089092255, + "learning_rate": 0.0006, + "loss": 6.658882141113281, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.07671099156141281, + "learning_rate": 0.0006, + "loss": 6.605809211730957, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.05295584350824356, + "learning_rate": 0.0006, + "loss": 6.641455173492432, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.09128042310476303, + "learning_rate": 0.0006, + "loss": 6.691066741943359, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.05362958461046219, + "learning_rate": 0.0006, + "loss": 6.573406219482422, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.046910062432289124, + "learning_rate": 0.0006, + "loss": 6.550876617431641, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.03494197130203247, + "learning_rate": 0.0006, + "loss": 6.540179252624512, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.04716387391090393, + "learning_rate": 0.0006, + "loss": 6.490459442138672, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.04190506786108017, + "learning_rate": 0.0006, + "loss": 6.509925365447998, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.04749064892530441, + "learning_rate": 0.0006, + "loss": 6.478779315948486, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.028002185747027397, + "learning_rate": 0.0006, + "loss": 6.528904914855957, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.036473166197538376, + "learning_rate": 0.0006, + "loss": 6.549042701721191, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.02543315477669239, + "learning_rate": 0.0006, + "loss": 6.498875141143799, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.029906807467341423, + "learning_rate": 0.0006, + "loss": 6.453019142150879, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.027231857180595398, + "learning_rate": 0.0006, + "loss": 6.428829193115234, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.02506154775619507, + "learning_rate": 0.0006, + "loss": 6.420858383178711, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0269472673535347, + "learning_rate": 0.0006, + "loss": 6.484989643096924, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.023410487920045853, + "learning_rate": 0.0006, + "loss": 6.454830169677734, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.034175023436546326, + "learning_rate": 0.0006, + "loss": 6.430975437164307, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 6.458896160125732, + "eval_runtime": 56.6934, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.045031629502773285, + "learning_rate": 0.0006, + "loss": 6.4243011474609375, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.04968734085559845, + "learning_rate": 0.0006, + "loss": 6.380091667175293, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.047145187854766846, + "learning_rate": 0.0006, + "loss": 6.477178573608398, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.04304948449134827, + "learning_rate": 0.0006, + "loss": 6.399726867675781, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.04466010630130768, + "learning_rate": 0.0006, + "loss": 6.3415679931640625, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.04305306077003479, + "learning_rate": 0.0006, + "loss": 6.3413310050964355, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.043064575642347336, + "learning_rate": 0.0006, + "loss": 6.4838714599609375, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.049503277987241745, + "learning_rate": 0.0006, + "loss": 6.38473653793335, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.04613238573074341, + "learning_rate": 0.0006, + "loss": 6.370769500732422, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.052119478583335876, + "learning_rate": 0.0006, + "loss": 6.362321853637695, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.045110076665878296, + "learning_rate": 0.0006, + "loss": 6.369844913482666, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.03798583149909973, + "learning_rate": 0.0006, + "loss": 6.324633598327637, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.03744659945368767, + "learning_rate": 0.0006, + "loss": 6.349421501159668, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.03767292574048042, + "learning_rate": 0.0006, + "loss": 6.334238052368164, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.03577509894967079, + "learning_rate": 0.0006, + "loss": 6.3453826904296875, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.03403240442276001, + "learning_rate": 0.0006, + "loss": 6.2836174964904785, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.03517331928014755, + "learning_rate": 0.0006, + "loss": 6.281718730926514, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.04409139230847359, + "learning_rate": 0.0006, + "loss": 6.320815086364746, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.054750364273786545, + "learning_rate": 0.0006, + "loss": 6.343618392944336, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.07612067461013794, + "learning_rate": 0.0006, + "loss": 6.310614585876465, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.08438985049724579, + "learning_rate": 0.0006, + "loss": 6.2369184494018555, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.0782201811671257, + "learning_rate": 0.0006, + "loss": 6.254452705383301, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.09928355365991592, + "learning_rate": 0.0006, + "loss": 6.3112077713012695, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.07233133912086487, + "learning_rate": 0.0006, + "loss": 6.408087253570557, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.07478684186935425, + "learning_rate": 0.0006, + "loss": 6.3770036697387695, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.06036846712231636, + "learning_rate": 0.0006, + "loss": 6.33865213394165, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.05783458054065704, + "learning_rate": 0.0006, + "loss": 6.363203048706055, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.05443909019231796, + "learning_rate": 0.0006, + "loss": 6.287069797515869, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.0557909794151783, + "learning_rate": 0.0006, + "loss": 6.297784805297852, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.03483372926712036, + "learning_rate": 0.0006, + "loss": 6.280734062194824, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.0425543375313282, + "learning_rate": 0.0006, + "loss": 6.33565616607666, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.03230159357190132, + "learning_rate": 0.0006, + "loss": 6.197698593139648, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.036897290498018265, + "learning_rate": 0.0006, + "loss": 6.241239547729492, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.031103838235139847, + "learning_rate": 0.0006, + "loss": 6.159435749053955, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.027116449549794197, + "learning_rate": 0.0006, + "loss": 6.212225914001465, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.029923981055617332, + "learning_rate": 0.0006, + "loss": 6.157269477844238, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.02521538734436035, + "learning_rate": 0.0006, + "loss": 6.146918773651123, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.02758920192718506, + "learning_rate": 0.0006, + "loss": 6.0446977615356445, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.024328220635652542, + "learning_rate": 0.0006, + "loss": 6.099508285522461, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.027425771579146385, + "learning_rate": 0.0006, + "loss": 6.084299087524414, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.026927651837468147, + "learning_rate": 0.0006, + "loss": 6.146334171295166, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.04054379090666771, + "learning_rate": 0.0006, + "loss": 6.1301493644714355, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.04456290975213051, + "learning_rate": 0.0006, + "loss": 6.152511119842529, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.052821505814790726, + "learning_rate": 0.0006, + "loss": 6.113544940948486, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.07251739501953125, + "learning_rate": 0.0006, + "loss": 6.118298530578613, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.09433556348085403, + "learning_rate": 0.0006, + "loss": 6.170897483825684, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.057683493942022324, + "learning_rate": 0.0006, + "loss": 6.1747822761535645, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.06363270431756973, + "learning_rate": 0.0006, + "loss": 6.1617889404296875, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.07528169453144073, + "learning_rate": 0.0006, + "loss": 6.230060577392578, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.07816170156002045, + "learning_rate": 0.0006, + "loss": 6.171257019042969, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.05860164389014244, + "learning_rate": 0.0006, + "loss": 6.095484256744385, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.04747429117560387, + "learning_rate": 0.0006, + "loss": 6.060067176818848, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.04514474794268608, + "learning_rate": 0.0006, + "loss": 6.100218772888184, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.05168681964278221, + "learning_rate": 0.0006, + "loss": 6.0621867179870605, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.04665876924991608, + "learning_rate": 0.0006, + "loss": 6.118292808532715, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.049374762922525406, + "learning_rate": 0.0006, + "loss": 6.098394393920898, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.03793549910187721, + "learning_rate": 0.0006, + "loss": 6.069541931152344, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.036362383514642715, + "learning_rate": 0.0006, + "loss": 6.1188435554504395, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.03397097811102867, + "learning_rate": 0.0006, + "loss": 6.117589950561523, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.030977753922343254, + "learning_rate": 0.0006, + "loss": 6.0236711502075195, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.02800217643380165, + "learning_rate": 0.0006, + "loss": 6.034825325012207, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.03129778802394867, + "learning_rate": 0.0006, + "loss": 5.980271339416504, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.025886094197630882, + "learning_rate": 0.0006, + "loss": 5.966419219970703, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.025712328031659126, + "learning_rate": 0.0006, + "loss": 5.938652515411377, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.022508298978209496, + "learning_rate": 0.0006, + "loss": 5.982312202453613, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.024880951270461082, + "learning_rate": 0.0006, + "loss": 5.931150436401367, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.030007576569914818, + "learning_rate": 0.0006, + "loss": 5.936639308929443, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.033314041793346405, + "learning_rate": 0.0006, + "loss": 5.944248676300049, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.043078381568193436, + "learning_rate": 0.0006, + "loss": 5.971286773681641, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.059296440333127975, + "learning_rate": 0.0006, + "loss": 5.910497665405273, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.06232666224241257, + "learning_rate": 0.0006, + "loss": 5.929817199707031, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.05303940549492836, + "learning_rate": 0.0006, + "loss": 5.910822868347168, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 5.977851390838623, + "eval_runtime": 56.2435, + "eval_samples_per_second": 43.418, + "eval_steps_per_second": 1.369, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.06943798810243607, + "learning_rate": 0.0006, + "loss": 5.854922294616699, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.05236278846859932, + "learning_rate": 0.0006, + "loss": 5.876611232757568, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.06240001320838928, + "learning_rate": 0.0006, + "loss": 5.878912448883057, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.07401192933320999, + "learning_rate": 0.0006, + "loss": 5.985136032104492, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.0826529711484909, + "learning_rate": 0.0006, + "loss": 5.939439296722412, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.07999005168676376, + "learning_rate": 0.0006, + "loss": 5.982336044311523, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.07014869153499603, + "learning_rate": 0.0006, + "loss": 5.896534442901611, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.05795128643512726, + "learning_rate": 0.0006, + "loss": 5.843111991882324, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.06374258548021317, + "learning_rate": 0.0006, + "loss": 5.78752326965332, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.045524805784225464, + "learning_rate": 0.0006, + "loss": 5.86755895614624, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.03773504123091698, + "learning_rate": 0.0006, + "loss": 5.83627462387085, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.035815201699733734, + "learning_rate": 0.0006, + "loss": 5.798238277435303, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.03756783902645111, + "learning_rate": 0.0006, + "loss": 5.845197677612305, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.03569113463163376, + "learning_rate": 0.0006, + "loss": 5.888932704925537, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.029467813670635223, + "learning_rate": 0.0006, + "loss": 5.792047500610352, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.025698406621813774, + "learning_rate": 0.0006, + "loss": 5.783734321594238, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.028426671400666237, + "learning_rate": 0.0006, + "loss": 5.762084007263184, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.03099622018635273, + "learning_rate": 0.0006, + "loss": 5.719109058380127, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.0310219656676054, + "learning_rate": 0.0006, + "loss": 5.737057209014893, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.02713446132838726, + "learning_rate": 0.0006, + "loss": 5.758657455444336, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.03099844604730606, + "learning_rate": 0.0006, + "loss": 5.742086410522461, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.030258730053901672, + "learning_rate": 0.0006, + "loss": 5.767426490783691, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.029034852981567383, + "learning_rate": 0.0006, + "loss": 5.651792049407959, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.03138665854930878, + "learning_rate": 0.0006, + "loss": 5.664980888366699, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.040546901524066925, + "learning_rate": 0.0006, + "loss": 5.665165901184082, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.06395485252141953, + "learning_rate": 0.0006, + "loss": 5.7776713371276855, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.10088056325912476, + "learning_rate": 0.0006, + "loss": 5.755645751953125, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.09527938067913055, + "learning_rate": 0.0006, + "loss": 5.838809490203857, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.0849648267030716, + "learning_rate": 0.0006, + "loss": 5.781932830810547, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.10142666101455688, + "learning_rate": 0.0006, + "loss": 5.838970184326172, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.42731454968452454, + "learning_rate": 0.0006, + "loss": 6.071902275085449, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.13102488219738007, + "learning_rate": 0.0006, + "loss": 6.8635993003845215, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.23691938817501068, + "learning_rate": 0.0006, + "loss": 7.313158988952637, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.11670669168233871, + "learning_rate": 0.0006, + "loss": 7.240363121032715, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.09985820949077606, + "learning_rate": 0.0006, + "loss": 7.113332271575928, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.11406408250331879, + "learning_rate": 0.0006, + "loss": 7.051285743713379, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.07900336384773254, + "learning_rate": 0.0006, + "loss": 7.014365196228027, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.05380333960056305, + "learning_rate": 0.0006, + "loss": 6.8868818283081055, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.05868449807167053, + "learning_rate": 0.0006, + "loss": 6.783599853515625, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.04428785294294357, + "learning_rate": 0.0006, + "loss": 6.798511505126953, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.055527277290821075, + "learning_rate": 0.0006, + "loss": 6.694507598876953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.03872432932257652, + "learning_rate": 0.0006, + "loss": 6.736631393432617, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.04330779239535332, + "learning_rate": 0.0006, + "loss": 6.576806545257568, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.036767736077308655, + "learning_rate": 0.0006, + "loss": 6.559871196746826, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.039857108145952225, + "learning_rate": 0.0006, + "loss": 6.5749831199646, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.048719972372055054, + "learning_rate": 0.0006, + "loss": 6.4915618896484375, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.042849015444517136, + "learning_rate": 0.0006, + "loss": 6.413516044616699, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.02871009334921837, + "learning_rate": 0.0006, + "loss": 6.320952415466309, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.03193620592355728, + "learning_rate": 0.0006, + "loss": 6.4176435470581055, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.027417024597525597, + "learning_rate": 0.0006, + "loss": 6.342319488525391, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.03305937349796295, + "learning_rate": 0.0006, + "loss": 6.246112823486328, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.031403183937072754, + "learning_rate": 0.0006, + "loss": 6.149503231048584, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.031563788652420044, + "learning_rate": 0.0006, + "loss": 6.278808116912842, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.028366589918732643, + "learning_rate": 0.0006, + "loss": 6.057036876678467, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.023715803399682045, + "learning_rate": 0.0006, + "loss": 6.111324310302734, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.026546811684966087, + "learning_rate": 0.0006, + "loss": 6.105669975280762, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.023531019687652588, + "learning_rate": 0.0006, + "loss": 6.068072319030762, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.024480128660798073, + "learning_rate": 0.0006, + "loss": 5.947896957397461, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.02517310529947281, + "learning_rate": 0.0006, + "loss": 6.030037879943848, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.026723403483629227, + "learning_rate": 0.0006, + "loss": 5.963776111602783, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.03163504600524902, + "learning_rate": 0.0006, + "loss": 5.893466949462891, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.04058683291077614, + "learning_rate": 0.0006, + "loss": 5.9202799797058105, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.04964117333292961, + "learning_rate": 0.0006, + "loss": 5.9366559982299805, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.055406033992767334, + "learning_rate": 0.0006, + "loss": 5.912230968475342, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.04643625393509865, + "learning_rate": 0.0006, + "loss": 5.87260103225708, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.029325876384973526, + "learning_rate": 0.0006, + "loss": 5.854763984680176, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.03159382566809654, + "learning_rate": 0.0006, + "loss": 5.880882263183594, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.030224012210965157, + "learning_rate": 0.0006, + "loss": 5.787360191345215, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.028368130326271057, + "learning_rate": 0.0006, + "loss": 5.817195892333984, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.029411399737000465, + "learning_rate": 0.0006, + "loss": 5.783403396606445, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.029927462339401245, + "learning_rate": 0.0006, + "loss": 5.662444591522217, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.02872961387038231, + "learning_rate": 0.0006, + "loss": 5.742781639099121, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 5.870260238647461, + "eval_runtime": 57.3734, + "eval_samples_per_second": 42.563, + "eval_steps_per_second": 1.342, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.028542131185531616, + "learning_rate": 0.0006, + "loss": 5.753256797790527, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.03227119892835617, + "learning_rate": 0.0006, + "loss": 5.722246170043945, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.031461723148822784, + "learning_rate": 0.0006, + "loss": 5.740520000457764, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.03184220567345619, + "learning_rate": 0.0006, + "loss": 5.720890045166016, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.030015580356121063, + "learning_rate": 0.0006, + "loss": 5.597814559936523, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.029931779950857162, + "learning_rate": 0.0006, + "loss": 5.5412468910217285, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.03206505626440048, + "learning_rate": 0.0006, + "loss": 5.6967597007751465, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.03978161886334419, + "learning_rate": 0.0006, + "loss": 5.59661340713501, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.04619080200791359, + "learning_rate": 0.0006, + "loss": 5.795788764953613, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.049085382372140884, + "learning_rate": 0.0006, + "loss": 5.667335510253906, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.04830858111381531, + "learning_rate": 0.0006, + "loss": 5.668806552886963, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0583178736269474, + "learning_rate": 0.0006, + "loss": 5.692285537719727, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.05182253196835518, + "learning_rate": 0.0006, + "loss": 5.678645133972168, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.04011905938386917, + "learning_rate": 0.0006, + "loss": 5.661550998687744, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.03888486698269844, + "learning_rate": 0.0006, + "loss": 5.643001556396484, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.030658265575766563, + "learning_rate": 0.0006, + "loss": 5.659090995788574, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.03609256446361542, + "learning_rate": 0.0006, + "loss": 5.544501781463623, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.033589769154787064, + "learning_rate": 0.0006, + "loss": 5.463878631591797, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.035128287971019745, + "learning_rate": 0.0006, + "loss": 5.5680060386657715, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.038684550672769547, + "learning_rate": 0.0006, + "loss": 5.587898254394531, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.043566226959228516, + "learning_rate": 0.0006, + "loss": 5.580896854400635, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.04428374394774437, + "learning_rate": 0.0006, + "loss": 5.603460311889648, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.037368934601545334, + "learning_rate": 0.0006, + "loss": 5.564740180969238, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.04092350974678993, + "learning_rate": 0.0006, + "loss": 5.584153652191162, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.045031506568193436, + "learning_rate": 0.0006, + "loss": 5.601186275482178, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.041830409318208694, + "learning_rate": 0.0006, + "loss": 5.512788772583008, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.037633974105119705, + "learning_rate": 0.0006, + "loss": 5.503828525543213, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.03473588079214096, + "learning_rate": 0.0006, + "loss": 5.570557594299316, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.039325181394815445, + "learning_rate": 0.0006, + "loss": 5.427783966064453, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.04034624621272087, + "learning_rate": 0.0006, + "loss": 5.572591304779053, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.05245184153318405, + "learning_rate": 0.0006, + "loss": 5.505208969116211, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.0568675734102726, + "learning_rate": 0.0006, + "loss": 5.589373588562012, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.048336587846279144, + "learning_rate": 0.0006, + "loss": 5.522005081176758, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.04194975271821022, + "learning_rate": 0.0006, + "loss": 5.387576103210449, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.04019433259963989, + "learning_rate": 0.0006, + "loss": 5.522615432739258, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.05709933117032051, + "learning_rate": 0.0006, + "loss": 5.420286178588867, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.09130338579416275, + "learning_rate": 0.0006, + "loss": 5.484622478485107, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.2757541835308075, + "learning_rate": 0.0006, + "loss": 5.821832656860352, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.28292080760002136, + "learning_rate": 0.0006, + "loss": 6.579934120178223, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.15293604135513306, + "learning_rate": 0.0006, + "loss": 6.478213310241699, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.11284393817186356, + "learning_rate": 0.0006, + "loss": 6.4066386222839355, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.10933073610067368, + "learning_rate": 0.0006, + "loss": 6.3306169509887695, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.05837641656398773, + "learning_rate": 0.0006, + "loss": 6.250000953674316, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.05987229570746422, + "learning_rate": 0.0006, + "loss": 6.152463436126709, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.04064938426017761, + "learning_rate": 0.0006, + "loss": 6.117897033691406, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.04565057530999184, + "learning_rate": 0.0006, + "loss": 6.042291641235352, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.03475063294172287, + "learning_rate": 0.0006, + "loss": 6.0071868896484375, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.03809220343828201, + "learning_rate": 0.0006, + "loss": 6.0128021240234375, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.03307219594717026, + "learning_rate": 0.0006, + "loss": 5.894664287567139, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.028824346140027046, + "learning_rate": 0.0006, + "loss": 5.914865493774414, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.027923690155148506, + "learning_rate": 0.0006, + "loss": 5.851759910583496, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.025492684915661812, + "learning_rate": 0.0006, + "loss": 5.749186038970947, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.025105174630880356, + "learning_rate": 0.0006, + "loss": 5.803566932678223, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.76276159286499, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.02468576282262802, + "learning_rate": 0.0006, + "loss": 5.833852767944336, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.022730689495801926, + "learning_rate": 0.0006, + "loss": 5.736724376678467, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.021335354074835777, + "learning_rate": 0.0006, + "loss": 5.644248008728027, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.02288101427257061, + "learning_rate": 0.0006, + "loss": 5.7011942863464355, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.01961585320532322, + "learning_rate": 0.0006, + "loss": 5.63215446472168, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.017659205943346024, + "learning_rate": 0.0006, + "loss": 5.744845390319824, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.02281784638762474, + "learning_rate": 0.0006, + "loss": 5.696993350982666, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.016394242644309998, + "learning_rate": 0.0006, + "loss": 5.659870147705078, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.01967368647456169, + "learning_rate": 0.0006, + "loss": 5.638072490692139, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.017326267436146736, + "learning_rate": 0.0006, + "loss": 5.638836860656738, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.017778463661670685, + "learning_rate": 0.0006, + "loss": 5.634466648101807, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.017269758507609367, + "learning_rate": 0.0006, + "loss": 5.506277561187744, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.01733289286494255, + "learning_rate": 0.0006, + "loss": 5.544455528259277, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.017787795513868332, + "learning_rate": 0.0006, + "loss": 5.498943328857422, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.015770932659506798, + "learning_rate": 0.0006, + "loss": 5.537227630615234, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.016132809221744537, + "learning_rate": 0.0006, + "loss": 5.489081382751465, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.01749664731323719, + "learning_rate": 0.0006, + "loss": 5.454263687133789, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.016711296513676643, + "learning_rate": 0.0006, + "loss": 5.530783653259277, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 5.596925258636475, + "eval_runtime": 56.9612, + "eval_samples_per_second": 42.871, + "eval_steps_per_second": 1.352, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.01867641694843769, + "learning_rate": 0.0006, + "loss": 5.506853103637695, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.023250792175531387, + "learning_rate": 0.0006, + "loss": 5.4074177742004395, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.01970936730504036, + "learning_rate": 0.0006, + "loss": 5.508777618408203, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.01879260502755642, + "learning_rate": 0.0006, + "loss": 5.412424087524414, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.017420092597603798, + "learning_rate": 0.0006, + "loss": 5.43361759185791, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.018707862123847008, + "learning_rate": 0.0006, + "loss": 5.476800918579102, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.023185882717370987, + "learning_rate": 0.0006, + "loss": 5.465272426605225, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.023570368066430092, + "learning_rate": 0.0006, + "loss": 5.391471862792969, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.02651144750416279, + "learning_rate": 0.0006, + "loss": 5.422799110412598, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.03316329047083855, + "learning_rate": 0.0006, + "loss": 5.3952202796936035, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.04012278467416763, + "learning_rate": 0.0006, + "loss": 5.540853023529053, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.04763203114271164, + "learning_rate": 0.0006, + "loss": 5.387890338897705, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.05129165202379227, + "learning_rate": 0.0006, + "loss": 5.414722442626953, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.04438071325421333, + "learning_rate": 0.0006, + "loss": 5.386244773864746, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.040234919637441635, + "learning_rate": 0.0006, + "loss": 5.431617736816406, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.03217008337378502, + "learning_rate": 0.0006, + "loss": 5.30168342590332, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.03087974339723587, + "learning_rate": 0.0006, + "loss": 5.316389083862305, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.02789347991347313, + "learning_rate": 0.0006, + "loss": 5.385970592498779, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.024792097508907318, + "learning_rate": 0.0006, + "loss": 5.362329959869385, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.025869522243738174, + "learning_rate": 0.0006, + "loss": 5.28400182723999, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.021339882165193558, + "learning_rate": 0.0006, + "loss": 5.36669921875, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.021462520584464073, + "learning_rate": 0.0006, + "loss": 5.311222076416016, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.02193879708647728, + "learning_rate": 0.0006, + "loss": 5.291441917419434, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.01963093690574169, + "learning_rate": 0.0006, + "loss": 5.342780113220215, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.020634736865758896, + "learning_rate": 0.0006, + "loss": 5.144141674041748, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.02126968279480934, + "learning_rate": 0.0006, + "loss": 5.232624530792236, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.021203994750976562, + "learning_rate": 0.0006, + "loss": 5.290219783782959, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.02193225920200348, + "learning_rate": 0.0006, + "loss": 5.29197359085083, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.023174211382865906, + "learning_rate": 0.0006, + "loss": 5.204349517822266, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.024161914363503456, + "learning_rate": 0.0006, + "loss": 5.339351654052734, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.02221524342894554, + "learning_rate": 0.0006, + "loss": 5.263001441955566, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.026632830500602722, + "learning_rate": 0.0006, + "loss": 5.221281051635742, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.034923870116472244, + "learning_rate": 0.0006, + "loss": 5.296039581298828, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.046184979379177094, + "learning_rate": 0.0006, + "loss": 5.212769985198975, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.05154098570346832, + "learning_rate": 0.0006, + "loss": 5.337864875793457, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.042967259883880615, + "learning_rate": 0.0006, + "loss": 5.267958641052246, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.05356583371758461, + "learning_rate": 0.0006, + "loss": 5.1404218673706055, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.08219032734632492, + "learning_rate": 0.0006, + "loss": 5.300687789916992, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.19204089045524597, + "learning_rate": 0.0006, + "loss": 5.456025123596191, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.25917696952819824, + "learning_rate": 0.0006, + "loss": 6.156761169433594, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.1566283404827118, + "learning_rate": 0.0006, + "loss": 5.862346649169922, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.1761486828327179, + "learning_rate": 0.0006, + "loss": 6.012795448303223, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.09857621043920517, + "learning_rate": 0.0006, + "loss": 5.9248151779174805, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.05968055501580238, + "learning_rate": 0.0006, + "loss": 5.780706882476807, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.057433564215898514, + "learning_rate": 0.0006, + "loss": 5.645388603210449, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.039659615606069565, + "learning_rate": 0.0006, + "loss": 5.695138931274414, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.04019695147871971, + "learning_rate": 0.0006, + "loss": 5.672435283660889, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.03033018670976162, + "learning_rate": 0.0006, + "loss": 5.682260513305664, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.03215787187218666, + "learning_rate": 0.0006, + "loss": 5.604182243347168, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.026865126565098763, + "learning_rate": 0.0006, + "loss": 5.64040470123291, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.02831609919667244, + "learning_rate": 0.0006, + "loss": 5.552720069885254, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.024226613342761993, + "learning_rate": 0.0006, + "loss": 5.556946277618408, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.025393934920430183, + "learning_rate": 0.0006, + "loss": 5.476178169250488, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.01984412968158722, + "learning_rate": 0.0006, + "loss": 5.471088409423828, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.02035129815340042, + "learning_rate": 0.0006, + "loss": 5.438812732696533, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0208350270986557, + "learning_rate": 0.0006, + "loss": 5.381746292114258, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.017783882096409798, + "learning_rate": 0.0006, + "loss": 5.351675510406494, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.01671770215034485, + "learning_rate": 0.0006, + "loss": 5.329428672790527, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.017891215160489082, + "learning_rate": 0.0006, + "loss": 5.212074279785156, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.016847791150212288, + "learning_rate": 0.0006, + "loss": 5.299534797668457, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.01593637838959694, + "learning_rate": 0.0006, + "loss": 5.30466890335083, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.014713102020323277, + "learning_rate": 0.0006, + "loss": 5.294345378875732, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.01433857623487711, + "learning_rate": 0.0006, + "loss": 5.2148542404174805, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.014979379251599312, + "learning_rate": 0.0006, + "loss": 5.298582077026367, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.013494192622601986, + "learning_rate": 0.0006, + "loss": 5.309783935546875, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.014020012691617012, + "learning_rate": 0.0006, + "loss": 5.284879684448242, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.01348058134317398, + "learning_rate": 0.0006, + "loss": 5.249116897583008, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.013280218467116356, + "learning_rate": 0.0006, + "loss": 5.273489952087402, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.012647110037505627, + "learning_rate": 0.0006, + "loss": 5.306652069091797, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.012847724370658398, + "learning_rate": 0.0006, + "loss": 5.235745429992676, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.011820166371762753, + "learning_rate": 0.0006, + "loss": 5.238091468811035, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.01255585066974163, + "learning_rate": 0.0006, + "loss": 5.231295585632324, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 5.394587993621826, + "eval_runtime": 56.3734, + "eval_samples_per_second": 43.318, + "eval_steps_per_second": 1.366, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.011031972244381905, + "learning_rate": 0.0006, + "loss": 5.1354570388793945, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.010631263256072998, + "learning_rate": 0.0006, + "loss": 5.177639961242676, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.01133775431662798, + "learning_rate": 0.0006, + "loss": 5.283428192138672, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.01154172234237194, + "learning_rate": 0.0006, + "loss": 5.206364154815674, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.010192356072366238, + "learning_rate": 0.0006, + "loss": 5.2385735511779785, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.012789200060069561, + "learning_rate": 0.0006, + "loss": 5.071800231933594, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.011282820254564285, + "learning_rate": 0.0006, + "loss": 5.218020439147949, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.012153448536992073, + "learning_rate": 0.0006, + "loss": 5.134698867797852, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.011920543387532234, + "learning_rate": 0.0006, + "loss": 5.141620635986328, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.010949329473078251, + "learning_rate": 0.0006, + "loss": 5.116259574890137, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.012362448498606682, + "learning_rate": 0.0006, + "loss": 5.090177059173584, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.011575071141123772, + "learning_rate": 0.0006, + "loss": 5.093441009521484, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.011126236990094185, + "learning_rate": 0.0006, + "loss": 5.060235023498535, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.0107401292771101, + "learning_rate": 0.0006, + "loss": 5.140704154968262, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.011202906258404255, + "learning_rate": 0.0006, + "loss": 5.10855770111084, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.011083454824984074, + "learning_rate": 0.0006, + "loss": 5.092751502990723, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.009523800574243069, + "learning_rate": 0.0006, + "loss": 5.13733434677124, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.011298353783786297, + "learning_rate": 0.0006, + "loss": 5.036048889160156, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.010676483623683453, + "learning_rate": 0.0006, + "loss": 5.051901340484619, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.011463399045169353, + "learning_rate": 0.0006, + "loss": 4.983170986175537, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.010603402741253376, + "learning_rate": 0.0006, + "loss": 5.112311363220215, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.010535670444369316, + "learning_rate": 0.0006, + "loss": 5.085227966308594, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.011843928135931492, + "learning_rate": 0.0006, + "loss": 5.07575798034668, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.010587952099740505, + "learning_rate": 0.0006, + "loss": 5.019237041473389, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.012285659089684486, + "learning_rate": 0.0006, + "loss": 5.144603729248047, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.013119431212544441, + "learning_rate": 0.0006, + "loss": 5.035256862640381, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.012138537131249905, + "learning_rate": 0.0006, + "loss": 4.996667861938477, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.012142804451286793, + "learning_rate": 0.0006, + "loss": 4.95058536529541, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.012205676175653934, + "learning_rate": 0.0006, + "loss": 5.096931457519531, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.012086744420230389, + "learning_rate": 0.0006, + "loss": 4.99384880065918, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.012918103486299515, + "learning_rate": 0.0006, + "loss": 5.031294822692871, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.012006713077425957, + "learning_rate": 0.0006, + "loss": 5.0547943115234375, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.0120399734005332, + "learning_rate": 0.0006, + "loss": 5.014238357543945, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.013306990265846252, + "learning_rate": 0.0006, + "loss": 4.972925186157227, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.013878636993467808, + "learning_rate": 0.0006, + "loss": 5.079033851623535, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.012454566545784473, + "learning_rate": 0.0006, + "loss": 4.93708610534668, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.014075061306357384, + "learning_rate": 0.0006, + "loss": 4.965582370758057, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.012525617145001888, + "learning_rate": 0.0006, + "loss": 4.968148708343506, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.012418225407600403, + "learning_rate": 0.0006, + "loss": 5.06901741027832, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.015034239739179611, + "learning_rate": 0.0006, + "loss": 4.984277725219727, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.014329628087580204, + "learning_rate": 0.0006, + "loss": 4.951558589935303, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.017769070342183113, + "learning_rate": 0.0006, + "loss": 4.966532230377197, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.01909453794360161, + "learning_rate": 0.0006, + "loss": 4.919890880584717, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.017519688233733177, + "learning_rate": 0.0006, + "loss": 5.031286716461182, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.017143897712230682, + "learning_rate": 0.0006, + "loss": 4.969764709472656, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.01688285917043686, + "learning_rate": 0.0006, + "loss": 5.0733442306518555, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.018547311425209045, + "learning_rate": 0.0006, + "loss": 4.980863571166992, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.018877137452363968, + "learning_rate": 0.0006, + "loss": 4.954646110534668, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.018681600689888, + "learning_rate": 0.0006, + "loss": 4.959702014923096, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.018947336822748184, + "learning_rate": 0.0006, + "loss": 5.070502758026123, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.018817484378814697, + "learning_rate": 0.0006, + "loss": 5.056508541107178, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.01931724324822426, + "learning_rate": 0.0006, + "loss": 4.922451972961426, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.022047819569706917, + "learning_rate": 0.0006, + "loss": 5.053170204162598, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.02279801294207573, + "learning_rate": 0.0006, + "loss": 4.8768439292907715, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.020661601796746254, + "learning_rate": 0.0006, + "loss": 4.907928466796875, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.020479509606957436, + "learning_rate": 0.0006, + "loss": 5.040694236755371, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.021654162555933, + "learning_rate": 0.0006, + "loss": 5.018434524536133, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.022486824542284012, + "learning_rate": 0.0006, + "loss": 5.086538314819336, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.02141369879245758, + "learning_rate": 0.0006, + "loss": 5.035694122314453, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.015926677733659744, + "learning_rate": 0.0006, + "loss": 4.905038833618164, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.015775298699736595, + "learning_rate": 0.0006, + "loss": 5.003902435302734, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.01872118003666401, + "learning_rate": 0.0006, + "loss": 4.891060829162598, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.02066192403435707, + "learning_rate": 0.0006, + "loss": 5.002897262573242, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.02348714880645275, + "learning_rate": 0.0006, + "loss": 4.9189958572387695, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.020402103662490845, + "learning_rate": 0.0006, + "loss": 4.992661476135254, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.020466072484850883, + "learning_rate": 0.0006, + "loss": 4.962968826293945, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.01922314241528511, + "learning_rate": 0.0006, + "loss": 4.897730827331543, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.017228754237294197, + "learning_rate": 0.0006, + "loss": 4.964842796325684, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.016021743416786194, + "learning_rate": 0.0006, + "loss": 4.910758972167969, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.017985103651881218, + "learning_rate": 0.0006, + "loss": 4.916072368621826, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.018780240789055824, + "learning_rate": 0.0006, + "loss": 4.94658088684082, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.01776314526796341, + "learning_rate": 0.0006, + "loss": 4.969357967376709, + "step": 3600 + }, + { + "epoch": 50.0, + "eval_loss": 5.141158580780029, + "eval_runtime": 56.2332, + "eval_samples_per_second": 43.426, + "eval_steps_per_second": 1.369, + "step": 3600 + }, + { + "epoch": 50.01397379912664, + "grad_norm": 0.016308696940541267, + "learning_rate": 0.0006, + "loss": 4.924664497375488, + "step": 3601 + }, + { + "epoch": 50.02794759825328, + "grad_norm": 0.018810199573636055, + "learning_rate": 0.0006, + "loss": 4.82293701171875, + "step": 3602 + }, + { + "epoch": 50.041921397379916, + "grad_norm": 0.02227160707116127, + "learning_rate": 0.0006, + "loss": 4.9612298011779785, + "step": 3603 + }, + { + "epoch": 50.05589519650655, + "grad_norm": 0.025266462936997414, + "learning_rate": 0.0006, + "loss": 4.963835716247559, + "step": 3604 + }, + { + "epoch": 50.069868995633186, + "grad_norm": 0.028588181361556053, + "learning_rate": 0.0006, + "loss": 4.992118835449219, + "step": 3605 + }, + { + "epoch": 50.083842794759825, + "grad_norm": 0.02739487960934639, + "learning_rate": 0.0006, + "loss": 4.915547847747803, + "step": 3606 + }, + { + "epoch": 50.09781659388646, + "grad_norm": 0.027000300586223602, + "learning_rate": 0.0006, + "loss": 4.883801460266113, + "step": 3607 + }, + { + "epoch": 50.1117903930131, + "grad_norm": 0.023468483239412308, + "learning_rate": 0.0006, + "loss": 5.006168365478516, + "step": 3608 + }, + { + "epoch": 50.12576419213974, + "grad_norm": 0.026753634214401245, + "learning_rate": 0.0006, + "loss": 4.92526912689209, + "step": 3609 + }, + { + "epoch": 50.13973799126637, + "grad_norm": 0.026816733181476593, + "learning_rate": 0.0006, + "loss": 4.847546100616455, + "step": 3610 + }, + { + "epoch": 50.15371179039301, + "grad_norm": 0.027362391352653503, + "learning_rate": 0.0006, + "loss": 4.862548351287842, + "step": 3611 + }, + { + "epoch": 50.16768558951965, + "grad_norm": 0.03450845181941986, + "learning_rate": 0.0006, + "loss": 4.963202953338623, + "step": 3612 + }, + { + "epoch": 50.18165938864629, + "grad_norm": 0.03801408037543297, + "learning_rate": 0.0006, + "loss": 4.770240306854248, + "step": 3613 + }, + { + "epoch": 50.19563318777293, + "grad_norm": 0.03579140082001686, + "learning_rate": 0.0006, + "loss": 4.797981262207031, + "step": 3614 + }, + { + "epoch": 50.209606986899566, + "grad_norm": 0.029393760487437248, + "learning_rate": 0.0006, + "loss": 4.798581600189209, + "step": 3615 + }, + { + "epoch": 50.223580786026204, + "grad_norm": 0.026111142709851265, + "learning_rate": 0.0006, + "loss": 4.990712642669678, + "step": 3616 + }, + { + "epoch": 50.237554585152836, + "grad_norm": 0.02213387005031109, + "learning_rate": 0.0006, + "loss": 4.865842342376709, + "step": 3617 + }, + { + "epoch": 50.251528384279474, + "grad_norm": 0.025450684130191803, + "learning_rate": 0.0006, + "loss": 4.903746128082275, + "step": 3618 + }, + { + "epoch": 50.26550218340611, + "grad_norm": 0.020915117114782333, + "learning_rate": 0.0006, + "loss": 4.951366424560547, + "step": 3619 + }, + { + "epoch": 50.27947598253275, + "grad_norm": 0.020972849801182747, + "learning_rate": 0.0006, + "loss": 4.783311367034912, + "step": 3620 + }, + { + "epoch": 50.29344978165939, + "grad_norm": 0.016956273466348648, + "learning_rate": 0.0006, + "loss": 4.955767631530762, + "step": 3621 + }, + { + "epoch": 50.30742358078603, + "grad_norm": 0.018045252189040184, + "learning_rate": 0.0006, + "loss": 4.869322776794434, + "step": 3622 + }, + { + "epoch": 50.32139737991266, + "grad_norm": 0.018199188634753227, + "learning_rate": 0.0006, + "loss": 4.95680046081543, + "step": 3623 + }, + { + "epoch": 50.3353711790393, + "grad_norm": 0.016688082367181778, + "learning_rate": 0.0006, + "loss": 4.783535957336426, + "step": 3624 + }, + { + "epoch": 50.34934497816594, + "grad_norm": 0.017540352419018745, + "learning_rate": 0.0006, + "loss": 4.8236083984375, + "step": 3625 + }, + { + "epoch": 50.36331877729258, + "grad_norm": 0.018054187297821045, + "learning_rate": 0.0006, + "loss": 4.764655113220215, + "step": 3626 + }, + { + "epoch": 50.377292576419215, + "grad_norm": 0.01741131953895092, + "learning_rate": 0.0006, + "loss": 4.875613689422607, + "step": 3627 + }, + { + "epoch": 50.391266375545854, + "grad_norm": 0.01682269014418125, + "learning_rate": 0.0006, + "loss": 4.818509101867676, + "step": 3628 + }, + { + "epoch": 50.40524017467249, + "grad_norm": 0.016652341932058334, + "learning_rate": 0.0006, + "loss": 4.974020481109619, + "step": 3629 + }, + { + "epoch": 50.419213973799124, + "grad_norm": 0.02035282365977764, + "learning_rate": 0.0006, + "loss": 4.893037796020508, + "step": 3630 + }, + { + "epoch": 50.43318777292576, + "grad_norm": 0.019907061010599136, + "learning_rate": 0.0006, + "loss": 4.848119735717773, + "step": 3631 + }, + { + "epoch": 50.4471615720524, + "grad_norm": 0.023861493915319443, + "learning_rate": 0.0006, + "loss": 4.819353103637695, + "step": 3632 + }, + { + "epoch": 50.46113537117904, + "grad_norm": 0.025254026055336, + "learning_rate": 0.0006, + "loss": 4.769312858581543, + "step": 3633 + }, + { + "epoch": 50.47510917030568, + "grad_norm": 0.026154937222599983, + "learning_rate": 0.0006, + "loss": 4.7586774826049805, + "step": 3634 + }, + { + "epoch": 50.48908296943232, + "grad_norm": 0.026342090219259262, + "learning_rate": 0.0006, + "loss": 4.861218452453613, + "step": 3635 + }, + { + "epoch": 50.50305676855895, + "grad_norm": 0.028169559314846992, + "learning_rate": 0.0006, + "loss": 4.818380355834961, + "step": 3636 + }, + { + "epoch": 50.51703056768559, + "grad_norm": 0.02820264920592308, + "learning_rate": 0.0006, + "loss": 4.892871856689453, + "step": 3637 + }, + { + "epoch": 50.531004366812226, + "grad_norm": 0.03345080465078354, + "learning_rate": 0.0006, + "loss": 4.811084747314453, + "step": 3638 + }, + { + "epoch": 50.544978165938865, + "grad_norm": 0.04005736857652664, + "learning_rate": 0.0006, + "loss": 4.765410900115967, + "step": 3639 + }, + { + "epoch": 50.5589519650655, + "grad_norm": 0.04333870857954025, + "learning_rate": 0.0006, + "loss": 4.956520080566406, + "step": 3640 + }, + { + "epoch": 50.57292576419214, + "grad_norm": 0.04216952994465828, + "learning_rate": 0.0006, + "loss": 4.793441295623779, + "step": 3641 + }, + { + "epoch": 50.58689956331878, + "grad_norm": 0.02936951443552971, + "learning_rate": 0.0006, + "loss": 4.8447957038879395, + "step": 3642 + }, + { + "epoch": 50.60087336244541, + "grad_norm": 0.026937026530504227, + "learning_rate": 0.0006, + "loss": 4.962215423583984, + "step": 3643 + }, + { + "epoch": 50.61484716157205, + "grad_norm": 0.02628890983760357, + "learning_rate": 0.0006, + "loss": 4.883061408996582, + "step": 3644 + }, + { + "epoch": 50.62882096069869, + "grad_norm": 0.025984834879636765, + "learning_rate": 0.0006, + "loss": 4.768662452697754, + "step": 3645 + }, + { + "epoch": 50.64279475982533, + "grad_norm": 0.023998012766242027, + "learning_rate": 0.0006, + "loss": 4.851325035095215, + "step": 3646 + }, + { + "epoch": 50.65676855895197, + "grad_norm": 0.02143184095621109, + "learning_rate": 0.0006, + "loss": 4.8274688720703125, + "step": 3647 + }, + { + "epoch": 50.670742358078606, + "grad_norm": 0.02488900162279606, + "learning_rate": 0.0006, + "loss": 4.831140518188477, + "step": 3648 + }, + { + "epoch": 50.68471615720524, + "grad_norm": 0.02330533228814602, + "learning_rate": 0.0006, + "loss": 4.850462913513184, + "step": 3649 + }, + { + "epoch": 50.698689956331876, + "grad_norm": 0.02121943049132824, + "learning_rate": 0.0006, + "loss": 4.870253086090088, + "step": 3650 + }, + { + "epoch": 50.712663755458514, + "grad_norm": 0.02080935798585415, + "learning_rate": 0.0006, + "loss": 4.798770904541016, + "step": 3651 + }, + { + "epoch": 50.72663755458515, + "grad_norm": 0.019047029316425323, + "learning_rate": 0.0006, + "loss": 4.8427300453186035, + "step": 3652 + }, + { + "epoch": 50.74061135371179, + "grad_norm": 0.01824103854596615, + "learning_rate": 0.0006, + "loss": 4.820980548858643, + "step": 3653 + }, + { + "epoch": 50.75458515283843, + "grad_norm": 0.018129967153072357, + "learning_rate": 0.0006, + "loss": 4.946732997894287, + "step": 3654 + }, + { + "epoch": 50.76855895196506, + "grad_norm": 0.01784815452992916, + "learning_rate": 0.0006, + "loss": 4.829516410827637, + "step": 3655 + }, + { + "epoch": 50.7825327510917, + "grad_norm": 0.020021555945277214, + "learning_rate": 0.0006, + "loss": 4.780428409576416, + "step": 3656 + }, + { + "epoch": 50.79650655021834, + "grad_norm": 0.01863253302872181, + "learning_rate": 0.0006, + "loss": 4.791567802429199, + "step": 3657 + }, + { + "epoch": 50.81048034934498, + "grad_norm": 0.019151683896780014, + "learning_rate": 0.0006, + "loss": 4.824892997741699, + "step": 3658 + }, + { + "epoch": 50.82445414847162, + "grad_norm": 0.01891648955643177, + "learning_rate": 0.0006, + "loss": 4.789605140686035, + "step": 3659 + }, + { + "epoch": 50.838427947598255, + "grad_norm": 0.016019510105252266, + "learning_rate": 0.0006, + "loss": 4.779555320739746, + "step": 3660 + }, + { + "epoch": 50.852401746724894, + "grad_norm": 0.016096491366624832, + "learning_rate": 0.0006, + "loss": 4.753842353820801, + "step": 3661 + }, + { + "epoch": 50.866375545851525, + "grad_norm": 0.01693476364016533, + "learning_rate": 0.0006, + "loss": 4.768477439880371, + "step": 3662 + }, + { + "epoch": 50.880349344978164, + "grad_norm": 0.01749575324356556, + "learning_rate": 0.0006, + "loss": 4.779143333435059, + "step": 3663 + }, + { + "epoch": 50.8943231441048, + "grad_norm": 0.016851043328642845, + "learning_rate": 0.0006, + "loss": 4.780387878417969, + "step": 3664 + }, + { + "epoch": 50.90829694323144, + "grad_norm": 0.019045202061533928, + "learning_rate": 0.0006, + "loss": 4.769913673400879, + "step": 3665 + }, + { + "epoch": 50.92227074235808, + "grad_norm": 0.01954123005270958, + "learning_rate": 0.0006, + "loss": 4.769009590148926, + "step": 3666 + }, + { + "epoch": 50.93624454148472, + "grad_norm": 0.022527096793055534, + "learning_rate": 0.0006, + "loss": 4.789215087890625, + "step": 3667 + }, + { + "epoch": 50.95021834061135, + "grad_norm": 0.02871634066104889, + "learning_rate": 0.0006, + "loss": 4.749727725982666, + "step": 3668 + }, + { + "epoch": 50.96419213973799, + "grad_norm": 0.04045053571462631, + "learning_rate": 0.0006, + "loss": 4.909038543701172, + "step": 3669 + }, + { + "epoch": 50.97816593886463, + "grad_norm": 0.04156879335641861, + "learning_rate": 0.0006, + "loss": 4.876831531524658, + "step": 3670 + }, + { + "epoch": 50.992139737991266, + "grad_norm": 0.02621397376060486, + "learning_rate": 0.0006, + "loss": 4.797173500061035, + "step": 3671 + }, + { + "epoch": 51.0, + "grad_norm": 0.03389597684144974, + "learning_rate": 0.0006, + "loss": 4.8428850173950195, + "step": 3672 + }, + { + "epoch": 51.0, + "eval_loss": 4.989261627197266, + "eval_runtime": 56.358, + "eval_samples_per_second": 43.33, + "eval_steps_per_second": 1.366, + "step": 3672 + }, + { + "epoch": 51.01397379912664, + "grad_norm": 0.03216972202062607, + "learning_rate": 0.0006, + "loss": 4.755102634429932, + "step": 3673 + }, + { + "epoch": 51.02794759825328, + "grad_norm": 0.03190374746918678, + "learning_rate": 0.0006, + "loss": 4.720424652099609, + "step": 3674 + }, + { + "epoch": 51.041921397379916, + "grad_norm": 0.031727299094200134, + "learning_rate": 0.0006, + "loss": 4.803277969360352, + "step": 3675 + }, + { + "epoch": 51.05589519650655, + "grad_norm": 0.02473079040646553, + "learning_rate": 0.0006, + "loss": 4.761087417602539, + "step": 3676 + }, + { + "epoch": 51.069868995633186, + "grad_norm": 0.020360209047794342, + "learning_rate": 0.0006, + "loss": 4.753884792327881, + "step": 3677 + }, + { + "epoch": 51.083842794759825, + "grad_norm": 0.02078923024237156, + "learning_rate": 0.0006, + "loss": 4.647130489349365, + "step": 3678 + }, + { + "epoch": 51.09781659388646, + "grad_norm": 0.019387200474739075, + "learning_rate": 0.0006, + "loss": 4.74135160446167, + "step": 3679 + }, + { + "epoch": 51.1117903930131, + "grad_norm": 0.01874949410557747, + "learning_rate": 0.0006, + "loss": 4.795329570770264, + "step": 3680 + }, + { + "epoch": 51.12576419213974, + "grad_norm": 0.019331837072968483, + "learning_rate": 0.0006, + "loss": 4.897967338562012, + "step": 3681 + }, + { + "epoch": 51.13973799126637, + "grad_norm": 0.017066050320863724, + "learning_rate": 0.0006, + "loss": 4.825351715087891, + "step": 3682 + }, + { + "epoch": 51.15371179039301, + "grad_norm": 0.01757810264825821, + "learning_rate": 0.0006, + "loss": 4.719335079193115, + "step": 3683 + }, + { + "epoch": 51.16768558951965, + "grad_norm": 0.018174266442656517, + "learning_rate": 0.0006, + "loss": 4.732680320739746, + "step": 3684 + }, + { + "epoch": 51.18165938864629, + "grad_norm": 0.018497727811336517, + "learning_rate": 0.0006, + "loss": 4.731050491333008, + "step": 3685 + }, + { + "epoch": 51.19563318777293, + "grad_norm": 0.0186338871717453, + "learning_rate": 0.0006, + "loss": 4.748035430908203, + "step": 3686 + }, + { + "epoch": 51.209606986899566, + "grad_norm": 0.018531711772084236, + "learning_rate": 0.0006, + "loss": 4.9007463455200195, + "step": 3687 + }, + { + "epoch": 51.223580786026204, + "grad_norm": 0.018221354112029076, + "learning_rate": 0.0006, + "loss": 4.84077262878418, + "step": 3688 + }, + { + "epoch": 51.237554585152836, + "grad_norm": 0.0203656405210495, + "learning_rate": 0.0006, + "loss": 4.73009729385376, + "step": 3689 + }, + { + "epoch": 51.251528384279474, + "grad_norm": 0.01964893378317356, + "learning_rate": 0.0006, + "loss": 4.772583961486816, + "step": 3690 + }, + { + "epoch": 51.26550218340611, + "grad_norm": 0.019606510177254677, + "learning_rate": 0.0006, + "loss": 4.599812030792236, + "step": 3691 + }, + { + "epoch": 51.27947598253275, + "grad_norm": 0.020501255989074707, + "learning_rate": 0.0006, + "loss": 4.809898853302002, + "step": 3692 + }, + { + "epoch": 51.29344978165939, + "grad_norm": 0.01870838552713394, + "learning_rate": 0.0006, + "loss": 4.741898059844971, + "step": 3693 + }, + { + "epoch": 51.30742358078603, + "grad_norm": 0.01802559196949005, + "learning_rate": 0.0006, + "loss": 4.746562480926514, + "step": 3694 + }, + { + "epoch": 51.32139737991266, + "grad_norm": 0.02053048275411129, + "learning_rate": 0.0006, + "loss": 4.69932746887207, + "step": 3695 + }, + { + "epoch": 51.3353711790393, + "grad_norm": 0.02661825157701969, + "learning_rate": 0.0006, + "loss": 4.68748664855957, + "step": 3696 + }, + { + "epoch": 51.34934497816594, + "grad_norm": 0.028269147500395775, + "learning_rate": 0.0006, + "loss": 4.731565475463867, + "step": 3697 + }, + { + "epoch": 51.36331877729258, + "grad_norm": 0.02650964818894863, + "learning_rate": 0.0006, + "loss": 4.733857154846191, + "step": 3698 + }, + { + "epoch": 51.377292576419215, + "grad_norm": 0.023696066811680794, + "learning_rate": 0.0006, + "loss": 4.812362194061279, + "step": 3699 + }, + { + "epoch": 51.391266375545854, + "grad_norm": 0.02365880273282528, + "learning_rate": 0.0006, + "loss": 4.733426094055176, + "step": 3700 + }, + { + "epoch": 51.40524017467249, + "grad_norm": 0.026249447837471962, + "learning_rate": 0.0006, + "loss": 4.715622901916504, + "step": 3701 + }, + { + "epoch": 51.419213973799124, + "grad_norm": 0.03301899880170822, + "learning_rate": 0.0006, + "loss": 4.7223663330078125, + "step": 3702 + }, + { + "epoch": 51.43318777292576, + "grad_norm": 0.044686660170555115, + "learning_rate": 0.0006, + "loss": 4.8121161460876465, + "step": 3703 + }, + { + "epoch": 51.4471615720524, + "grad_norm": 0.04733522981405258, + "learning_rate": 0.0006, + "loss": 4.841394424438477, + "step": 3704 + }, + { + "epoch": 51.46113537117904, + "grad_norm": 0.0395209975540638, + "learning_rate": 0.0006, + "loss": 4.773837089538574, + "step": 3705 + }, + { + "epoch": 51.47510917030568, + "grad_norm": 0.032790958881378174, + "learning_rate": 0.0006, + "loss": 4.833066463470459, + "step": 3706 + }, + { + "epoch": 51.48908296943232, + "grad_norm": 0.030578427016735077, + "learning_rate": 0.0006, + "loss": 4.685190200805664, + "step": 3707 + }, + { + "epoch": 51.50305676855895, + "grad_norm": 0.031619083136320114, + "learning_rate": 0.0006, + "loss": 4.740378379821777, + "step": 3708 + }, + { + "epoch": 51.51703056768559, + "grad_norm": 0.02385825850069523, + "learning_rate": 0.0006, + "loss": 4.812911510467529, + "step": 3709 + }, + { + "epoch": 51.531004366812226, + "grad_norm": 0.023523425683379173, + "learning_rate": 0.0006, + "loss": 4.669580459594727, + "step": 3710 + }, + { + "epoch": 51.544978165938865, + "grad_norm": 0.021316280588507652, + "learning_rate": 0.0006, + "loss": 4.76888370513916, + "step": 3711 + }, + { + "epoch": 51.5589519650655, + "grad_norm": 0.024663880467414856, + "learning_rate": 0.0006, + "loss": 4.702592372894287, + "step": 3712 + }, + { + "epoch": 51.57292576419214, + "grad_norm": 0.024173280224204063, + "learning_rate": 0.0006, + "loss": 4.710604667663574, + "step": 3713 + }, + { + "epoch": 51.58689956331878, + "grad_norm": 0.02754114381968975, + "learning_rate": 0.0006, + "loss": 4.746356964111328, + "step": 3714 + }, + { + "epoch": 51.60087336244541, + "grad_norm": 0.027903622016310692, + "learning_rate": 0.0006, + "loss": 4.7546491622924805, + "step": 3715 + }, + { + "epoch": 51.61484716157205, + "grad_norm": 0.024871617555618286, + "learning_rate": 0.0006, + "loss": 4.605746269226074, + "step": 3716 + }, + { + "epoch": 51.62882096069869, + "grad_norm": 0.021239031106233597, + "learning_rate": 0.0006, + "loss": 4.6942458152771, + "step": 3717 + }, + { + "epoch": 51.64279475982533, + "grad_norm": 0.01925622671842575, + "learning_rate": 0.0006, + "loss": 4.593558311462402, + "step": 3718 + }, + { + "epoch": 51.65676855895197, + "grad_norm": 0.01956283673644066, + "learning_rate": 0.0006, + "loss": 4.631172180175781, + "step": 3719 + }, + { + "epoch": 51.670742358078606, + "grad_norm": 0.02269871160387993, + "learning_rate": 0.0006, + "loss": 4.729236602783203, + "step": 3720 + }, + { + "epoch": 51.68471615720524, + "grad_norm": 0.02253916673362255, + "learning_rate": 0.0006, + "loss": 4.7191033363342285, + "step": 3721 + }, + { + "epoch": 51.698689956331876, + "grad_norm": 0.026231050491333008, + "learning_rate": 0.0006, + "loss": 4.811850547790527, + "step": 3722 + }, + { + "epoch": 51.712663755458514, + "grad_norm": 0.02558821812272072, + "learning_rate": 0.0006, + "loss": 4.681466102600098, + "step": 3723 + }, + { + "epoch": 51.72663755458515, + "grad_norm": 0.02124209702014923, + "learning_rate": 0.0006, + "loss": 4.693664073944092, + "step": 3724 + }, + { + "epoch": 51.74061135371179, + "grad_norm": 0.016638319939374924, + "learning_rate": 0.0006, + "loss": 4.771440029144287, + "step": 3725 + }, + { + "epoch": 51.75458515283843, + "grad_norm": 0.018048347905278206, + "learning_rate": 0.0006, + "loss": 4.643252372741699, + "step": 3726 + }, + { + "epoch": 51.76855895196506, + "grad_norm": 0.019904717803001404, + "learning_rate": 0.0006, + "loss": 4.66530704498291, + "step": 3727 + }, + { + "epoch": 51.7825327510917, + "grad_norm": 0.0194614976644516, + "learning_rate": 0.0006, + "loss": 4.711943626403809, + "step": 3728 + }, + { + "epoch": 51.79650655021834, + "grad_norm": 0.019175555557012558, + "learning_rate": 0.0006, + "loss": 4.721931457519531, + "step": 3729 + }, + { + "epoch": 51.81048034934498, + "grad_norm": 0.02093447372317314, + "learning_rate": 0.0006, + "loss": 4.633059024810791, + "step": 3730 + }, + { + "epoch": 51.82445414847162, + "grad_norm": 0.01930283196270466, + "learning_rate": 0.0006, + "loss": 4.659373760223389, + "step": 3731 + }, + { + "epoch": 51.838427947598255, + "grad_norm": 0.015350676141679287, + "learning_rate": 0.0006, + "loss": 4.763708114624023, + "step": 3732 + }, + { + "epoch": 51.852401746724894, + "grad_norm": 0.014139448292553425, + "learning_rate": 0.0006, + "loss": 4.753891468048096, + "step": 3733 + }, + { + "epoch": 51.866375545851525, + "grad_norm": 0.016215872019529343, + "learning_rate": 0.0006, + "loss": 4.658967018127441, + "step": 3734 + }, + { + "epoch": 51.880349344978164, + "grad_norm": 0.017328619956970215, + "learning_rate": 0.0006, + "loss": 4.696579933166504, + "step": 3735 + }, + { + "epoch": 51.8943231441048, + "grad_norm": 0.018103396520018578, + "learning_rate": 0.0006, + "loss": 4.627464771270752, + "step": 3736 + }, + { + "epoch": 51.90829694323144, + "grad_norm": 0.01715235412120819, + "learning_rate": 0.0006, + "loss": 4.7368011474609375, + "step": 3737 + }, + { + "epoch": 51.92227074235808, + "grad_norm": 0.014790094457566738, + "learning_rate": 0.0006, + "loss": 4.629441738128662, + "step": 3738 + }, + { + "epoch": 51.93624454148472, + "grad_norm": 0.017114296555519104, + "learning_rate": 0.0006, + "loss": 4.761256694793701, + "step": 3739 + }, + { + "epoch": 51.95021834061135, + "grad_norm": 0.019009845331311226, + "learning_rate": 0.0006, + "loss": 4.789155006408691, + "step": 3740 + }, + { + "epoch": 51.96419213973799, + "grad_norm": 0.019696692004799843, + "learning_rate": 0.0006, + "loss": 4.721269607543945, + "step": 3741 + }, + { + "epoch": 51.97816593886463, + "grad_norm": 0.020542822778224945, + "learning_rate": 0.0006, + "loss": 4.687038898468018, + "step": 3742 + }, + { + "epoch": 51.992139737991266, + "grad_norm": 0.021267160773277283, + "learning_rate": 0.0006, + "loss": 4.747860908508301, + "step": 3743 + }, + { + "epoch": 52.0, + "grad_norm": 0.025049567222595215, + "learning_rate": 0.0006, + "loss": 4.8473100662231445, + "step": 3744 + }, + { + "epoch": 52.0, + "eval_loss": 4.943043231964111, + "eval_runtime": 56.4929, + "eval_samples_per_second": 43.227, + "eval_steps_per_second": 1.363, + "step": 3744 + }, + { + "epoch": 52.01397379912664, + "grad_norm": 0.03054986707866192, + "learning_rate": 0.0006, + "loss": 4.856931686401367, + "step": 3745 + }, + { + "epoch": 52.02794759825328, + "grad_norm": 0.03838375583291054, + "learning_rate": 0.0006, + "loss": 4.699807167053223, + "step": 3746 + }, + { + "epoch": 52.041921397379916, + "grad_norm": 0.03383488208055496, + "learning_rate": 0.0006, + "loss": 4.6626787185668945, + "step": 3747 + }, + { + "epoch": 52.05589519650655, + "grad_norm": 0.0201561376452446, + "learning_rate": 0.0006, + "loss": 4.621831893920898, + "step": 3748 + }, + { + "epoch": 52.069868995633186, + "grad_norm": 0.028749438002705574, + "learning_rate": 0.0006, + "loss": 4.682540416717529, + "step": 3749 + }, + { + "epoch": 52.083842794759825, + "grad_norm": 0.02681174874305725, + "learning_rate": 0.0006, + "loss": 4.629997253417969, + "step": 3750 + }, + { + "epoch": 52.09781659388646, + "grad_norm": 0.024063318967819214, + "learning_rate": 0.0006, + "loss": 4.573021411895752, + "step": 3751 + }, + { + "epoch": 52.1117903930131, + "grad_norm": 0.02324330434203148, + "learning_rate": 0.0006, + "loss": 4.739465236663818, + "step": 3752 + }, + { + "epoch": 52.12576419213974, + "grad_norm": 0.0225426834076643, + "learning_rate": 0.0006, + "loss": 4.7335333824157715, + "step": 3753 + }, + { + "epoch": 52.13973799126637, + "grad_norm": 0.019593428820371628, + "learning_rate": 0.0006, + "loss": 4.704958438873291, + "step": 3754 + }, + { + "epoch": 52.15371179039301, + "grad_norm": 0.020777082070708275, + "learning_rate": 0.0006, + "loss": 4.638696670532227, + "step": 3755 + }, + { + "epoch": 52.16768558951965, + "grad_norm": 0.021525604650378227, + "learning_rate": 0.0006, + "loss": 4.643706798553467, + "step": 3756 + }, + { + "epoch": 52.18165938864629, + "grad_norm": 0.018943654373288155, + "learning_rate": 0.0006, + "loss": 4.5173211097717285, + "step": 3757 + }, + { + "epoch": 52.19563318777293, + "grad_norm": 0.01856580562889576, + "learning_rate": 0.0006, + "loss": 4.678788661956787, + "step": 3758 + }, + { + "epoch": 52.209606986899566, + "grad_norm": 0.019417259842157364, + "learning_rate": 0.0006, + "loss": 4.587244987487793, + "step": 3759 + }, + { + "epoch": 52.223580786026204, + "grad_norm": 0.01911911740899086, + "learning_rate": 0.0006, + "loss": 4.688503265380859, + "step": 3760 + }, + { + "epoch": 52.237554585152836, + "grad_norm": 0.021017879247665405, + "learning_rate": 0.0006, + "loss": 4.65829610824585, + "step": 3761 + }, + { + "epoch": 52.251528384279474, + "grad_norm": 0.02048875205218792, + "learning_rate": 0.0006, + "loss": 4.624055862426758, + "step": 3762 + }, + { + "epoch": 52.26550218340611, + "grad_norm": 0.01932772807776928, + "learning_rate": 0.0006, + "loss": 4.556562900543213, + "step": 3763 + }, + { + "epoch": 52.27947598253275, + "grad_norm": 0.01871614344418049, + "learning_rate": 0.0006, + "loss": 4.581201076507568, + "step": 3764 + }, + { + "epoch": 52.29344978165939, + "grad_norm": 0.017253106459975243, + "learning_rate": 0.0006, + "loss": 4.605219841003418, + "step": 3765 + }, + { + "epoch": 52.30742358078603, + "grad_norm": 0.01831238530576229, + "learning_rate": 0.0006, + "loss": 4.759900093078613, + "step": 3766 + }, + { + "epoch": 52.32139737991266, + "grad_norm": 0.017946207895874977, + "learning_rate": 0.0006, + "loss": 4.5598344802856445, + "step": 3767 + }, + { + "epoch": 52.3353711790393, + "grad_norm": 0.01680106669664383, + "learning_rate": 0.0006, + "loss": 4.605034828186035, + "step": 3768 + }, + { + "epoch": 52.34934497816594, + "grad_norm": 0.021482540294528008, + "learning_rate": 0.0006, + "loss": 4.620411396026611, + "step": 3769 + }, + { + "epoch": 52.36331877729258, + "grad_norm": 0.02210252732038498, + "learning_rate": 0.0006, + "loss": 4.667577743530273, + "step": 3770 + }, + { + "epoch": 52.377292576419215, + "grad_norm": 0.02237485907971859, + "learning_rate": 0.0006, + "loss": 4.628812789916992, + "step": 3771 + }, + { + "epoch": 52.391266375545854, + "grad_norm": 0.02196541801095009, + "learning_rate": 0.0006, + "loss": 4.572988510131836, + "step": 3772 + }, + { + "epoch": 52.40524017467249, + "grad_norm": 0.01930229924619198, + "learning_rate": 0.0006, + "loss": 4.588339805603027, + "step": 3773 + }, + { + "epoch": 52.419213973799124, + "grad_norm": 0.02222434990108013, + "learning_rate": 0.0006, + "loss": 4.766482353210449, + "step": 3774 + }, + { + "epoch": 52.43318777292576, + "grad_norm": 0.028806306421756744, + "learning_rate": 0.0006, + "loss": 4.617892742156982, + "step": 3775 + }, + { + "epoch": 52.4471615720524, + "grad_norm": 0.03306042402982712, + "learning_rate": 0.0006, + "loss": 4.665334701538086, + "step": 3776 + }, + { + "epoch": 52.46113537117904, + "grad_norm": 0.0358956903219223, + "learning_rate": 0.0006, + "loss": 4.566668510437012, + "step": 3777 + }, + { + "epoch": 52.47510917030568, + "grad_norm": 0.030406497418880463, + "learning_rate": 0.0006, + "loss": 4.513612270355225, + "step": 3778 + }, + { + "epoch": 52.48908296943232, + "grad_norm": 0.030295712873339653, + "learning_rate": 0.0006, + "loss": 4.640189170837402, + "step": 3779 + }, + { + "epoch": 52.50305676855895, + "grad_norm": 0.024981003254652023, + "learning_rate": 0.0006, + "loss": 4.57558012008667, + "step": 3780 + }, + { + "epoch": 52.51703056768559, + "grad_norm": 0.024990713223814964, + "learning_rate": 0.0006, + "loss": 4.708589553833008, + "step": 3781 + }, + { + "epoch": 52.531004366812226, + "grad_norm": 0.025034775957465172, + "learning_rate": 0.0006, + "loss": 4.70468282699585, + "step": 3782 + }, + { + "epoch": 52.544978165938865, + "grad_norm": 0.028243759647011757, + "learning_rate": 0.0006, + "loss": 4.712060451507568, + "step": 3783 + }, + { + "epoch": 52.5589519650655, + "grad_norm": 0.026984544470906258, + "learning_rate": 0.0006, + "loss": 4.505004405975342, + "step": 3784 + }, + { + "epoch": 52.57292576419214, + "grad_norm": 0.050843529403209686, + "learning_rate": 0.0006, + "loss": 4.593271255493164, + "step": 3785 + }, + { + "epoch": 52.58689956331878, + "grad_norm": 0.25422802567481995, + "learning_rate": 0.0006, + "loss": 4.7178802490234375, + "step": 3786 + }, + { + "epoch": 52.60087336244541, + "grad_norm": 0.6078406572341919, + "learning_rate": 0.0006, + "loss": 6.144372940063477, + "step": 3787 + }, + { + "epoch": 52.61484716157205, + "grad_norm": 0.25933101773262024, + "learning_rate": 0.0006, + "loss": 5.971083641052246, + "step": 3788 + }, + { + "epoch": 52.62882096069869, + "grad_norm": 0.294657826423645, + "learning_rate": 0.0006, + "loss": 6.204090118408203, + "step": 3789 + }, + { + "epoch": 52.64279475982533, + "grad_norm": 0.23641522228717804, + "learning_rate": 0.0006, + "loss": 6.869574546813965, + "step": 3790 + }, + { + "epoch": 52.65676855895197, + "grad_norm": 0.23717311024665833, + "learning_rate": 0.0006, + "loss": 6.739034175872803, + "step": 3791 + }, + { + "epoch": 52.670742358078606, + "grad_norm": 0.10300362855195999, + "learning_rate": 0.0006, + "loss": 6.49191951751709, + "step": 3792 + }, + { + "epoch": 52.68471615720524, + "grad_norm": 0.1484968513250351, + "learning_rate": 0.0006, + "loss": 6.6021270751953125, + "step": 3793 + }, + { + "epoch": 52.698689956331876, + "grad_norm": 0.08052115142345428, + "learning_rate": 0.0006, + "loss": 6.432796478271484, + "step": 3794 + }, + { + "epoch": 52.712663755458514, + "grad_norm": 0.06869253516197205, + "learning_rate": 0.0006, + "loss": 6.212784767150879, + "step": 3795 + }, + { + "epoch": 52.72663755458515, + "grad_norm": 0.05589582398533821, + "learning_rate": 0.0006, + "loss": 6.168288707733154, + "step": 3796 + }, + { + "epoch": 52.74061135371179, + "grad_norm": 0.04840940609574318, + "learning_rate": 0.0006, + "loss": 6.163174629211426, + "step": 3797 + }, + { + "epoch": 52.75458515283843, + "grad_norm": 0.04578601196408272, + "learning_rate": 0.0006, + "loss": 6.038390636444092, + "step": 3798 + }, + { + "epoch": 52.76855895196506, + "grad_norm": 0.04102053865790367, + "learning_rate": 0.0006, + "loss": 5.929539680480957, + "step": 3799 + }, + { + "epoch": 52.7825327510917, + "grad_norm": 0.038398485630750656, + "learning_rate": 0.0006, + "loss": 6.007378101348877, + "step": 3800 + }, + { + "epoch": 52.79650655021834, + "grad_norm": 0.03466642275452614, + "learning_rate": 0.0006, + "loss": 5.767291069030762, + "step": 3801 + }, + { + "epoch": 52.81048034934498, + "grad_norm": 0.03328905999660492, + "learning_rate": 0.0006, + "loss": 5.785032272338867, + "step": 3802 + }, + { + "epoch": 52.82445414847162, + "grad_norm": 0.03186681121587753, + "learning_rate": 0.0006, + "loss": 5.6942057609558105, + "step": 3803 + }, + { + "epoch": 52.838427947598255, + "grad_norm": 0.030553551390767097, + "learning_rate": 0.0006, + "loss": 5.680774688720703, + "step": 3804 + }, + { + "epoch": 52.852401746724894, + "grad_norm": 0.02578839473426342, + "learning_rate": 0.0006, + "loss": 5.703461647033691, + "step": 3805 + }, + { + "epoch": 52.866375545851525, + "grad_norm": 0.026261312887072563, + "learning_rate": 0.0006, + "loss": 5.642549514770508, + "step": 3806 + }, + { + "epoch": 52.880349344978164, + "grad_norm": 0.02782430872321129, + "learning_rate": 0.0006, + "loss": 5.563597679138184, + "step": 3807 + }, + { + "epoch": 52.8943231441048, + "grad_norm": 0.024467799812555313, + "learning_rate": 0.0006, + "loss": 5.583395957946777, + "step": 3808 + }, + { + "epoch": 52.90829694323144, + "grad_norm": 0.022602897137403488, + "learning_rate": 0.0006, + "loss": 5.50742769241333, + "step": 3809 + }, + { + "epoch": 52.92227074235808, + "grad_norm": 0.024292029440402985, + "learning_rate": 0.0006, + "loss": 5.367804527282715, + "step": 3810 + }, + { + "epoch": 52.93624454148472, + "grad_norm": 0.021062295883893967, + "learning_rate": 0.0006, + "loss": 5.423705101013184, + "step": 3811 + }, + { + "epoch": 52.95021834061135, + "grad_norm": 0.022226469591259956, + "learning_rate": 0.0006, + "loss": 5.344039440155029, + "step": 3812 + }, + { + "epoch": 52.96419213973799, + "grad_norm": 0.023691752925515175, + "learning_rate": 0.0006, + "loss": 5.2581892013549805, + "step": 3813 + }, + { + "epoch": 52.97816593886463, + "grad_norm": 0.02181999757885933, + "learning_rate": 0.0006, + "loss": 5.2710113525390625, + "step": 3814 + }, + { + "epoch": 52.992139737991266, + "grad_norm": 0.025295857340097427, + "learning_rate": 0.0006, + "loss": 5.278355598449707, + "step": 3815 + }, + { + "epoch": 53.0, + "grad_norm": 0.02216828241944313, + "learning_rate": 0.0006, + "loss": 5.127945899963379, + "step": 3816 + }, + { + "epoch": 53.0, + "eval_loss": 5.3411736488342285, + "eval_runtime": 56.2837, + "eval_samples_per_second": 43.387, + "eval_steps_per_second": 1.368, + "step": 3816 + }, + { + "epoch": 53.01397379912664, + "grad_norm": 0.022496026009321213, + "learning_rate": 0.0006, + "loss": 5.19655704498291, + "step": 3817 + }, + { + "epoch": 53.02794759825328, + "grad_norm": 0.03201168775558472, + "learning_rate": 0.0006, + "loss": 5.094621181488037, + "step": 3818 + }, + { + "epoch": 53.041921397379916, + "grad_norm": 0.02683178335428238, + "learning_rate": 0.0006, + "loss": 5.19899845123291, + "step": 3819 + }, + { + "epoch": 53.05589519650655, + "grad_norm": 0.022659169510006905, + "learning_rate": 0.0006, + "loss": 5.184103012084961, + "step": 3820 + }, + { + "epoch": 53.069868995633186, + "grad_norm": 0.024784443899989128, + "learning_rate": 0.0006, + "loss": 5.124286651611328, + "step": 3821 + }, + { + "epoch": 53.083842794759825, + "grad_norm": 0.028710370883345604, + "learning_rate": 0.0006, + "loss": 5.130426406860352, + "step": 3822 + }, + { + "epoch": 53.09781659388646, + "grad_norm": 0.02366088517010212, + "learning_rate": 0.0006, + "loss": 5.1453399658203125, + "step": 3823 + }, + { + "epoch": 53.1117903930131, + "grad_norm": 0.01813938096165657, + "learning_rate": 0.0006, + "loss": 5.05271053314209, + "step": 3824 + }, + { + "epoch": 53.12576419213974, + "grad_norm": 0.02051435597240925, + "learning_rate": 0.0006, + "loss": 5.023038864135742, + "step": 3825 + }, + { + "epoch": 53.13973799126637, + "grad_norm": 0.02185754105448723, + "learning_rate": 0.0006, + "loss": 4.8950018882751465, + "step": 3826 + }, + { + "epoch": 53.15371179039301, + "grad_norm": 0.017223743721842766, + "learning_rate": 0.0006, + "loss": 5.078621864318848, + "step": 3827 + }, + { + "epoch": 53.16768558951965, + "grad_norm": 0.01832217164337635, + "learning_rate": 0.0006, + "loss": 4.921341419219971, + "step": 3828 + }, + { + "epoch": 53.18165938864629, + "grad_norm": 0.02114005573093891, + "learning_rate": 0.0006, + "loss": 4.997865676879883, + "step": 3829 + }, + { + "epoch": 53.19563318777293, + "grad_norm": 0.021426431834697723, + "learning_rate": 0.0006, + "loss": 4.877785682678223, + "step": 3830 + }, + { + "epoch": 53.209606986899566, + "grad_norm": 0.021327560767531395, + "learning_rate": 0.0006, + "loss": 4.847512245178223, + "step": 3831 + }, + { + "epoch": 53.223580786026204, + "grad_norm": 0.027825530618429184, + "learning_rate": 0.0006, + "loss": 5.004823207855225, + "step": 3832 + }, + { + "epoch": 53.237554585152836, + "grad_norm": 0.03756481409072876, + "learning_rate": 0.0006, + "loss": 4.817290306091309, + "step": 3833 + }, + { + "epoch": 53.251528384279474, + "grad_norm": 0.04933721572160721, + "learning_rate": 0.0006, + "loss": 4.863317489624023, + "step": 3834 + }, + { + "epoch": 53.26550218340611, + "grad_norm": 0.04588694870471954, + "learning_rate": 0.0006, + "loss": 4.864742755889893, + "step": 3835 + }, + { + "epoch": 53.27947598253275, + "grad_norm": 0.02728264033794403, + "learning_rate": 0.0006, + "loss": 4.9068756103515625, + "step": 3836 + }, + { + "epoch": 53.29344978165939, + "grad_norm": 0.036240726709365845, + "learning_rate": 0.0006, + "loss": 4.902045249938965, + "step": 3837 + }, + { + "epoch": 53.30742358078603, + "grad_norm": 0.02703195810317993, + "learning_rate": 0.0006, + "loss": 4.908980369567871, + "step": 3838 + }, + { + "epoch": 53.32139737991266, + "grad_norm": 0.024894049391150475, + "learning_rate": 0.0006, + "loss": 4.794848442077637, + "step": 3839 + }, + { + "epoch": 53.3353711790393, + "grad_norm": 0.02389993891119957, + "learning_rate": 0.0006, + "loss": 4.8234734535217285, + "step": 3840 + }, + { + "epoch": 53.34934497816594, + "grad_norm": 0.02486400306224823, + "learning_rate": 0.0006, + "loss": 4.851463794708252, + "step": 3841 + }, + { + "epoch": 53.36331877729258, + "grad_norm": 0.02613813802599907, + "learning_rate": 0.0006, + "loss": 4.860444068908691, + "step": 3842 + }, + { + "epoch": 53.377292576419215, + "grad_norm": 0.0229744054377079, + "learning_rate": 0.0006, + "loss": 4.709606170654297, + "step": 3843 + }, + { + "epoch": 53.391266375545854, + "grad_norm": 0.02123042568564415, + "learning_rate": 0.0006, + "loss": 4.654613494873047, + "step": 3844 + }, + { + "epoch": 53.40524017467249, + "grad_norm": 0.017352338880300522, + "learning_rate": 0.0006, + "loss": 4.769761085510254, + "step": 3845 + }, + { + "epoch": 53.419213973799124, + "grad_norm": 0.019808724522590637, + "learning_rate": 0.0006, + "loss": 4.98514461517334, + "step": 3846 + }, + { + "epoch": 53.43318777292576, + "grad_norm": 0.018776152282953262, + "learning_rate": 0.0006, + "loss": 4.749990463256836, + "step": 3847 + }, + { + "epoch": 53.4471615720524, + "grad_norm": 0.02017521858215332, + "learning_rate": 0.0006, + "loss": 4.757381439208984, + "step": 3848 + }, + { + "epoch": 53.46113537117904, + "grad_norm": 0.019509416073560715, + "learning_rate": 0.0006, + "loss": 4.8111572265625, + "step": 3849 + }, + { + "epoch": 53.47510917030568, + "grad_norm": 0.015543580986559391, + "learning_rate": 0.0006, + "loss": 4.7501678466796875, + "step": 3850 + }, + { + "epoch": 53.48908296943232, + "grad_norm": 0.017941651865839958, + "learning_rate": 0.0006, + "loss": 4.701838493347168, + "step": 3851 + }, + { + "epoch": 53.50305676855895, + "grad_norm": 0.01678212359547615, + "learning_rate": 0.0006, + "loss": 4.787798881530762, + "step": 3852 + }, + { + "epoch": 53.51703056768559, + "grad_norm": 0.019050654023885727, + "learning_rate": 0.0006, + "loss": 4.754859924316406, + "step": 3853 + }, + { + "epoch": 53.531004366812226, + "grad_norm": 0.017406078055500984, + "learning_rate": 0.0006, + "loss": 4.752772808074951, + "step": 3854 + }, + { + "epoch": 53.544978165938865, + "grad_norm": 0.01773344725370407, + "learning_rate": 0.0006, + "loss": 4.651812553405762, + "step": 3855 + }, + { + "epoch": 53.5589519650655, + "grad_norm": 0.017221346497535706, + "learning_rate": 0.0006, + "loss": 4.792244911193848, + "step": 3856 + }, + { + "epoch": 53.57292576419214, + "grad_norm": 0.01722758449614048, + "learning_rate": 0.0006, + "loss": 4.7902960777282715, + "step": 3857 + }, + { + "epoch": 53.58689956331878, + "grad_norm": 0.020445192232728004, + "learning_rate": 0.0006, + "loss": 4.67392110824585, + "step": 3858 + }, + { + "epoch": 53.60087336244541, + "grad_norm": 0.02267114631831646, + "learning_rate": 0.0006, + "loss": 4.796089172363281, + "step": 3859 + }, + { + "epoch": 53.61484716157205, + "grad_norm": 0.027129942551255226, + "learning_rate": 0.0006, + "loss": 4.774303913116455, + "step": 3860 + }, + { + "epoch": 53.62882096069869, + "grad_norm": 0.025454839691519737, + "learning_rate": 0.0006, + "loss": 4.708813190460205, + "step": 3861 + }, + { + "epoch": 53.64279475982533, + "grad_norm": 0.019248757511377335, + "learning_rate": 0.0006, + "loss": 4.760467529296875, + "step": 3862 + }, + { + "epoch": 53.65676855895197, + "grad_norm": 0.01791176199913025, + "learning_rate": 0.0006, + "loss": 4.789203643798828, + "step": 3863 + }, + { + "epoch": 53.670742358078606, + "grad_norm": 0.016306612640619278, + "learning_rate": 0.0006, + "loss": 4.732605934143066, + "step": 3864 + }, + { + "epoch": 53.68471615720524, + "grad_norm": 0.017836976796388626, + "learning_rate": 0.0006, + "loss": 4.691871643066406, + "step": 3865 + }, + { + "epoch": 53.698689956331876, + "grad_norm": 0.015102861449122429, + "learning_rate": 0.0006, + "loss": 4.724156856536865, + "step": 3866 + }, + { + "epoch": 53.712663755458514, + "grad_norm": 0.017458142712712288, + "learning_rate": 0.0006, + "loss": 4.73524808883667, + "step": 3867 + }, + { + "epoch": 53.72663755458515, + "grad_norm": 0.01591932773590088, + "learning_rate": 0.0006, + "loss": 4.705888748168945, + "step": 3868 + }, + { + "epoch": 53.74061135371179, + "grad_norm": 0.01713593676686287, + "learning_rate": 0.0006, + "loss": 4.644516944885254, + "step": 3869 + }, + { + "epoch": 53.75458515283843, + "grad_norm": 0.01905885338783264, + "learning_rate": 0.0006, + "loss": 4.698416233062744, + "step": 3870 + }, + { + "epoch": 53.76855895196506, + "grad_norm": 0.01766437292098999, + "learning_rate": 0.0006, + "loss": 4.6796555519104, + "step": 3871 + }, + { + "epoch": 53.7825327510917, + "grad_norm": 0.01552243810147047, + "learning_rate": 0.0006, + "loss": 4.78138542175293, + "step": 3872 + }, + { + "epoch": 53.79650655021834, + "grad_norm": 0.014191331341862679, + "learning_rate": 0.0006, + "loss": 4.703588485717773, + "step": 3873 + }, + { + "epoch": 53.81048034934498, + "grad_norm": 0.015872279182076454, + "learning_rate": 0.0006, + "loss": 4.705611228942871, + "step": 3874 + }, + { + "epoch": 53.82445414847162, + "grad_norm": 0.01787560060620308, + "learning_rate": 0.0006, + "loss": 4.8227949142456055, + "step": 3875 + }, + { + "epoch": 53.838427947598255, + "grad_norm": 0.0158209390938282, + "learning_rate": 0.0006, + "loss": 4.645227432250977, + "step": 3876 + }, + { + "epoch": 53.852401746724894, + "grad_norm": 0.014362678863108158, + "learning_rate": 0.0006, + "loss": 4.7225422859191895, + "step": 3877 + }, + { + "epoch": 53.866375545851525, + "grad_norm": 0.014899161644279957, + "learning_rate": 0.0006, + "loss": 4.685914039611816, + "step": 3878 + }, + { + "epoch": 53.880349344978164, + "grad_norm": 0.01628505438566208, + "learning_rate": 0.0006, + "loss": 4.674065589904785, + "step": 3879 + }, + { + "epoch": 53.8943231441048, + "grad_norm": 0.01873331144452095, + "learning_rate": 0.0006, + "loss": 4.811744213104248, + "step": 3880 + }, + { + "epoch": 53.90829694323144, + "grad_norm": 0.02229069173336029, + "learning_rate": 0.0006, + "loss": 4.721514701843262, + "step": 3881 + }, + { + "epoch": 53.92227074235808, + "grad_norm": 0.028012916445732117, + "learning_rate": 0.0006, + "loss": 4.773603439331055, + "step": 3882 + }, + { + "epoch": 53.93624454148472, + "grad_norm": 0.036322783678770065, + "learning_rate": 0.0006, + "loss": 4.787134170532227, + "step": 3883 + }, + { + "epoch": 53.95021834061135, + "grad_norm": 0.048149142414331436, + "learning_rate": 0.0006, + "loss": 4.736196517944336, + "step": 3884 + }, + { + "epoch": 53.96419213973799, + "grad_norm": 0.051918916404247284, + "learning_rate": 0.0006, + "loss": 4.686728000640869, + "step": 3885 + }, + { + "epoch": 53.97816593886463, + "grad_norm": 0.0481342189013958, + "learning_rate": 0.0006, + "loss": 4.733730792999268, + "step": 3886 + }, + { + "epoch": 53.992139737991266, + "grad_norm": 0.045424818992614746, + "learning_rate": 0.0006, + "loss": 4.748723983764648, + "step": 3887 + }, + { + "epoch": 54.0, + "grad_norm": 0.04153316840529442, + "learning_rate": 0.0006, + "loss": 4.715220928192139, + "step": 3888 + }, + { + "epoch": 54.0, + "eval_loss": 4.964215278625488, + "eval_runtime": 56.7757, + "eval_samples_per_second": 43.011, + "eval_steps_per_second": 1.356, + "step": 3888 + }, + { + "epoch": 54.01397379912664, + "grad_norm": 0.03459750860929489, + "learning_rate": 0.0006, + "loss": 4.661557674407959, + "step": 3889 + }, + { + "epoch": 54.02794759825328, + "grad_norm": 0.025409208610653877, + "learning_rate": 0.0006, + "loss": 4.7065582275390625, + "step": 3890 + }, + { + "epoch": 54.041921397379916, + "grad_norm": 0.03202550485730171, + "learning_rate": 0.0006, + "loss": 4.751070976257324, + "step": 3891 + }, + { + "epoch": 54.05589519650655, + "grad_norm": 0.02968018501996994, + "learning_rate": 0.0006, + "loss": 4.635388374328613, + "step": 3892 + }, + { + "epoch": 54.069868995633186, + "grad_norm": 0.02757406234741211, + "learning_rate": 0.0006, + "loss": 4.706214427947998, + "step": 3893 + }, + { + "epoch": 54.083842794759825, + "grad_norm": 0.025223620235919952, + "learning_rate": 0.0006, + "loss": 4.703792572021484, + "step": 3894 + }, + { + "epoch": 54.09781659388646, + "grad_norm": 0.02431778982281685, + "learning_rate": 0.0006, + "loss": 4.658116340637207, + "step": 3895 + }, + { + "epoch": 54.1117903930131, + "grad_norm": 0.02792172320187092, + "learning_rate": 0.0006, + "loss": 4.638895034790039, + "step": 3896 + }, + { + "epoch": 54.12576419213974, + "grad_norm": 0.032253511250019073, + "learning_rate": 0.0006, + "loss": 4.693902492523193, + "step": 3897 + }, + { + "epoch": 54.13973799126637, + "grad_norm": 0.027044931426644325, + "learning_rate": 0.0006, + "loss": 4.575805187225342, + "step": 3898 + }, + { + "epoch": 54.15371179039301, + "grad_norm": 0.023270361125469208, + "learning_rate": 0.0006, + "loss": 4.67645263671875, + "step": 3899 + }, + { + "epoch": 54.16768558951965, + "grad_norm": 0.01929079182446003, + "learning_rate": 0.0006, + "loss": 4.68062686920166, + "step": 3900 + }, + { + "epoch": 54.18165938864629, + "grad_norm": 0.020343473181128502, + "learning_rate": 0.0006, + "loss": 4.708086967468262, + "step": 3901 + }, + { + "epoch": 54.19563318777293, + "grad_norm": 0.0211793202906847, + "learning_rate": 0.0006, + "loss": 4.677323341369629, + "step": 3902 + }, + { + "epoch": 54.209606986899566, + "grad_norm": 0.020341217517852783, + "learning_rate": 0.0006, + "loss": 4.760937690734863, + "step": 3903 + }, + { + "epoch": 54.223580786026204, + "grad_norm": 0.019421054050326347, + "learning_rate": 0.0006, + "loss": 4.508834362030029, + "step": 3904 + }, + { + "epoch": 54.237554585152836, + "grad_norm": 0.016701679676771164, + "learning_rate": 0.0006, + "loss": 4.664399147033691, + "step": 3905 + }, + { + "epoch": 54.251528384279474, + "grad_norm": 0.016561470925807953, + "learning_rate": 0.0006, + "loss": 4.742990493774414, + "step": 3906 + }, + { + "epoch": 54.26550218340611, + "grad_norm": 0.017446085810661316, + "learning_rate": 0.0006, + "loss": 4.63038444519043, + "step": 3907 + }, + { + "epoch": 54.27947598253275, + "grad_norm": 0.014340086840093136, + "learning_rate": 0.0006, + "loss": 4.688791275024414, + "step": 3908 + }, + { + "epoch": 54.29344978165939, + "grad_norm": 0.013580698519945145, + "learning_rate": 0.0006, + "loss": 4.625302314758301, + "step": 3909 + }, + { + "epoch": 54.30742358078603, + "grad_norm": 0.014931539073586464, + "learning_rate": 0.0006, + "loss": 4.6167497634887695, + "step": 3910 + }, + { + "epoch": 54.32139737991266, + "grad_norm": 0.013492150232195854, + "learning_rate": 0.0006, + "loss": 4.657521724700928, + "step": 3911 + }, + { + "epoch": 54.3353711790393, + "grad_norm": 0.015151413157582283, + "learning_rate": 0.0006, + "loss": 4.634521484375, + "step": 3912 + }, + { + "epoch": 54.34934497816594, + "grad_norm": 0.01302040833979845, + "learning_rate": 0.0006, + "loss": 4.718928337097168, + "step": 3913 + }, + { + "epoch": 54.36331877729258, + "grad_norm": 0.013433529995381832, + "learning_rate": 0.0006, + "loss": 4.584803581237793, + "step": 3914 + }, + { + "epoch": 54.377292576419215, + "grad_norm": 0.012987039051949978, + "learning_rate": 0.0006, + "loss": 4.6393585205078125, + "step": 3915 + }, + { + "epoch": 54.391266375545854, + "grad_norm": 0.013695075176656246, + "learning_rate": 0.0006, + "loss": 4.556313514709473, + "step": 3916 + }, + { + "epoch": 54.40524017467249, + "grad_norm": 0.013718255795538425, + "learning_rate": 0.0006, + "loss": 4.576223850250244, + "step": 3917 + }, + { + "epoch": 54.419213973799124, + "grad_norm": 0.013699112460017204, + "learning_rate": 0.0006, + "loss": 4.63889217376709, + "step": 3918 + }, + { + "epoch": 54.43318777292576, + "grad_norm": 0.013516264036297798, + "learning_rate": 0.0006, + "loss": 4.669532299041748, + "step": 3919 + }, + { + "epoch": 54.4471615720524, + "grad_norm": 0.013597294688224792, + "learning_rate": 0.0006, + "loss": 4.6644287109375, + "step": 3920 + }, + { + "epoch": 54.46113537117904, + "grad_norm": 0.013185849413275719, + "learning_rate": 0.0006, + "loss": 4.636900901794434, + "step": 3921 + }, + { + "epoch": 54.47510917030568, + "grad_norm": 0.014311990700662136, + "learning_rate": 0.0006, + "loss": 4.6047587394714355, + "step": 3922 + }, + { + "epoch": 54.48908296943232, + "grad_norm": 0.012426027096807957, + "learning_rate": 0.0006, + "loss": 4.6161394119262695, + "step": 3923 + }, + { + "epoch": 54.50305676855895, + "grad_norm": 0.01283493172377348, + "learning_rate": 0.0006, + "loss": 4.584190368652344, + "step": 3924 + }, + { + "epoch": 54.51703056768559, + "grad_norm": 0.01339943427592516, + "learning_rate": 0.0006, + "loss": 4.624184608459473, + "step": 3925 + }, + { + "epoch": 54.531004366812226, + "grad_norm": 0.012859724462032318, + "learning_rate": 0.0006, + "loss": 4.6174635887146, + "step": 3926 + }, + { + "epoch": 54.544978165938865, + "grad_norm": 0.014420023187994957, + "learning_rate": 0.0006, + "loss": 4.536544322967529, + "step": 3927 + }, + { + "epoch": 54.5589519650655, + "grad_norm": 0.014237185008823872, + "learning_rate": 0.0006, + "loss": 4.681514739990234, + "step": 3928 + }, + { + "epoch": 54.57292576419214, + "grad_norm": 0.013340328820049763, + "learning_rate": 0.0006, + "loss": 4.579689979553223, + "step": 3929 + }, + { + "epoch": 54.58689956331878, + "grad_norm": 0.013894530944526196, + "learning_rate": 0.0006, + "loss": 4.638958930969238, + "step": 3930 + }, + { + "epoch": 54.60087336244541, + "grad_norm": 0.016181141138076782, + "learning_rate": 0.0006, + "loss": 4.573508262634277, + "step": 3931 + }, + { + "epoch": 54.61484716157205, + "grad_norm": 0.013221992179751396, + "learning_rate": 0.0006, + "loss": 4.6229047775268555, + "step": 3932 + }, + { + "epoch": 54.62882096069869, + "grad_norm": 0.013241561129689217, + "learning_rate": 0.0006, + "loss": 4.641152858734131, + "step": 3933 + }, + { + "epoch": 54.64279475982533, + "grad_norm": 0.016237538307905197, + "learning_rate": 0.0006, + "loss": 4.6421217918396, + "step": 3934 + }, + { + "epoch": 54.65676855895197, + "grad_norm": 0.020866507664322853, + "learning_rate": 0.0006, + "loss": 4.646058082580566, + "step": 3935 + }, + { + "epoch": 54.670742358078606, + "grad_norm": 0.022182893007993698, + "learning_rate": 0.0006, + "loss": 4.554501533508301, + "step": 3936 + }, + { + "epoch": 54.68471615720524, + "grad_norm": 0.02254868485033512, + "learning_rate": 0.0006, + "loss": 4.563869953155518, + "step": 3937 + }, + { + "epoch": 54.698689956331876, + "grad_norm": 0.019730940461158752, + "learning_rate": 0.0006, + "loss": 4.592165946960449, + "step": 3938 + }, + { + "epoch": 54.712663755458514, + "grad_norm": 0.015146561898291111, + "learning_rate": 0.0006, + "loss": 4.6228742599487305, + "step": 3939 + }, + { + "epoch": 54.72663755458515, + "grad_norm": 0.01824716106057167, + "learning_rate": 0.0006, + "loss": 4.716236114501953, + "step": 3940 + }, + { + "epoch": 54.74061135371179, + "grad_norm": 0.017730550840497017, + "learning_rate": 0.0006, + "loss": 4.650232315063477, + "step": 3941 + }, + { + "epoch": 54.75458515283843, + "grad_norm": 0.014875595457851887, + "learning_rate": 0.0006, + "loss": 4.6377458572387695, + "step": 3942 + }, + { + "epoch": 54.76855895196506, + "grad_norm": 0.01613209955394268, + "learning_rate": 0.0006, + "loss": 4.5768537521362305, + "step": 3943 + }, + { + "epoch": 54.7825327510917, + "grad_norm": 0.017311280593276024, + "learning_rate": 0.0006, + "loss": 4.625744342803955, + "step": 3944 + }, + { + "epoch": 54.79650655021834, + "grad_norm": 0.018770085647702217, + "learning_rate": 0.0006, + "loss": 4.613787651062012, + "step": 3945 + }, + { + "epoch": 54.81048034934498, + "grad_norm": 0.0165257528424263, + "learning_rate": 0.0006, + "loss": 4.5991315841674805, + "step": 3946 + }, + { + "epoch": 54.82445414847162, + "grad_norm": 0.017433857545256615, + "learning_rate": 0.0006, + "loss": 4.513872146606445, + "step": 3947 + }, + { + "epoch": 54.838427947598255, + "grad_norm": 0.020126905292272568, + "learning_rate": 0.0006, + "loss": 4.616185188293457, + "step": 3948 + }, + { + "epoch": 54.852401746724894, + "grad_norm": 0.01819358579814434, + "learning_rate": 0.0006, + "loss": 4.535691261291504, + "step": 3949 + }, + { + "epoch": 54.866375545851525, + "grad_norm": 0.01899927295744419, + "learning_rate": 0.0006, + "loss": 4.595199108123779, + "step": 3950 + }, + { + "epoch": 54.880349344978164, + "grad_norm": 0.02005566656589508, + "learning_rate": 0.0006, + "loss": 4.633289337158203, + "step": 3951 + }, + { + "epoch": 54.8943231441048, + "grad_norm": 0.021570220589637756, + "learning_rate": 0.0006, + "loss": 4.592503547668457, + "step": 3952 + }, + { + "epoch": 54.90829694323144, + "grad_norm": 0.01883506216108799, + "learning_rate": 0.0006, + "loss": 4.6597795486450195, + "step": 3953 + }, + { + "epoch": 54.92227074235808, + "grad_norm": 0.018070021644234657, + "learning_rate": 0.0006, + "loss": 4.601474285125732, + "step": 3954 + }, + { + "epoch": 54.93624454148472, + "grad_norm": 0.018090378493070602, + "learning_rate": 0.0006, + "loss": 4.569846153259277, + "step": 3955 + }, + { + "epoch": 54.95021834061135, + "grad_norm": 0.018751563504338264, + "learning_rate": 0.0006, + "loss": 4.646064281463623, + "step": 3956 + }, + { + "epoch": 54.96419213973799, + "grad_norm": 0.017512254416942596, + "learning_rate": 0.0006, + "loss": 4.606670379638672, + "step": 3957 + }, + { + "epoch": 54.97816593886463, + "grad_norm": 0.018771110102534294, + "learning_rate": 0.0006, + "loss": 4.604315757751465, + "step": 3958 + }, + { + "epoch": 54.992139737991266, + "grad_norm": 0.01611141301691532, + "learning_rate": 0.0006, + "loss": 4.643245220184326, + "step": 3959 + }, + { + "epoch": 55.0, + "grad_norm": 0.0156728383153677, + "learning_rate": 0.0006, + "loss": 4.6739044189453125, + "step": 3960 + }, + { + "epoch": 55.0, + "eval_loss": 4.851032257080078, + "eval_runtime": 56.8699, + "eval_samples_per_second": 42.94, + "eval_steps_per_second": 1.354, + "step": 3960 + }, + { + "epoch": 55.01397379912664, + "grad_norm": 0.016214672476053238, + "learning_rate": 0.0006, + "loss": 4.571749687194824, + "step": 3961 + }, + { + "epoch": 55.02794759825328, + "grad_norm": 0.01649155467748642, + "learning_rate": 0.0006, + "loss": 4.553592205047607, + "step": 3962 + }, + { + "epoch": 55.041921397379916, + "grad_norm": 0.016793379560112953, + "learning_rate": 0.0006, + "loss": 4.650628566741943, + "step": 3963 + }, + { + "epoch": 55.05589519650655, + "grad_norm": 0.0191937405616045, + "learning_rate": 0.0006, + "loss": 4.620532989501953, + "step": 3964 + }, + { + "epoch": 55.069868995633186, + "grad_norm": 0.020976556465029716, + "learning_rate": 0.0006, + "loss": 4.6329545974731445, + "step": 3965 + }, + { + "epoch": 55.083842794759825, + "grad_norm": 0.023450477048754692, + "learning_rate": 0.0006, + "loss": 4.60828971862793, + "step": 3966 + }, + { + "epoch": 55.09781659388646, + "grad_norm": 0.026778312399983406, + "learning_rate": 0.0006, + "loss": 4.617481708526611, + "step": 3967 + }, + { + "epoch": 55.1117903930131, + "grad_norm": 0.02333568036556244, + "learning_rate": 0.0006, + "loss": 4.580722332000732, + "step": 3968 + }, + { + "epoch": 55.12576419213974, + "grad_norm": 0.018491310998797417, + "learning_rate": 0.0006, + "loss": 4.590953350067139, + "step": 3969 + }, + { + "epoch": 55.13973799126637, + "grad_norm": 0.020339803770184517, + "learning_rate": 0.0006, + "loss": 4.518374443054199, + "step": 3970 + }, + { + "epoch": 55.15371179039301, + "grad_norm": 0.01759706623852253, + "learning_rate": 0.0006, + "loss": 4.560522079467773, + "step": 3971 + }, + { + "epoch": 55.16768558951965, + "grad_norm": 0.015875980257987976, + "learning_rate": 0.0006, + "loss": 4.4795708656311035, + "step": 3972 + }, + { + "epoch": 55.18165938864629, + "grad_norm": 0.01650134287774563, + "learning_rate": 0.0006, + "loss": 4.57435941696167, + "step": 3973 + }, + { + "epoch": 55.19563318777293, + "grad_norm": 0.01734331250190735, + "learning_rate": 0.0006, + "loss": 4.661007881164551, + "step": 3974 + }, + { + "epoch": 55.209606986899566, + "grad_norm": 0.01651303470134735, + "learning_rate": 0.0006, + "loss": 4.511148452758789, + "step": 3975 + }, + { + "epoch": 55.223580786026204, + "grad_norm": 0.016189761459827423, + "learning_rate": 0.0006, + "loss": 4.592350006103516, + "step": 3976 + }, + { + "epoch": 55.237554585152836, + "grad_norm": 0.01833222061395645, + "learning_rate": 0.0006, + "loss": 4.506649971008301, + "step": 3977 + }, + { + "epoch": 55.251528384279474, + "grad_norm": 0.02433410845696926, + "learning_rate": 0.0006, + "loss": 4.49500036239624, + "step": 3978 + }, + { + "epoch": 55.26550218340611, + "grad_norm": 0.0314289927482605, + "learning_rate": 0.0006, + "loss": 4.507624626159668, + "step": 3979 + }, + { + "epoch": 55.27947598253275, + "grad_norm": 0.03013679012656212, + "learning_rate": 0.0006, + "loss": 4.533641815185547, + "step": 3980 + }, + { + "epoch": 55.29344978165939, + "grad_norm": 0.024524061009287834, + "learning_rate": 0.0006, + "loss": 4.5721116065979, + "step": 3981 + }, + { + "epoch": 55.30742358078603, + "grad_norm": 0.02640336938202381, + "learning_rate": 0.0006, + "loss": 4.652521133422852, + "step": 3982 + }, + { + "epoch": 55.32139737991266, + "grad_norm": 0.021309101954102516, + "learning_rate": 0.0006, + "loss": 4.600375175476074, + "step": 3983 + }, + { + "epoch": 55.3353711790393, + "grad_norm": 0.023008549585938454, + "learning_rate": 0.0006, + "loss": 4.53914737701416, + "step": 3984 + }, + { + "epoch": 55.34934497816594, + "grad_norm": 0.022491468116641045, + "learning_rate": 0.0006, + "loss": 4.661163330078125, + "step": 3985 + }, + { + "epoch": 55.36331877729258, + "grad_norm": 0.01904081553220749, + "learning_rate": 0.0006, + "loss": 4.5426177978515625, + "step": 3986 + }, + { + "epoch": 55.377292576419215, + "grad_norm": 0.019229518249630928, + "learning_rate": 0.0006, + "loss": 4.606066703796387, + "step": 3987 + }, + { + "epoch": 55.391266375545854, + "grad_norm": 0.01781383901834488, + "learning_rate": 0.0006, + "loss": 4.54642391204834, + "step": 3988 + }, + { + "epoch": 55.40524017467249, + "grad_norm": 0.0174600500613451, + "learning_rate": 0.0006, + "loss": 4.711891174316406, + "step": 3989 + }, + { + "epoch": 55.419213973799124, + "grad_norm": 0.018154339864850044, + "learning_rate": 0.0006, + "loss": 4.527122974395752, + "step": 3990 + }, + { + "epoch": 55.43318777292576, + "grad_norm": 0.01914648339152336, + "learning_rate": 0.0006, + "loss": 4.641975402832031, + "step": 3991 + }, + { + "epoch": 55.4471615720524, + "grad_norm": 0.017118625342845917, + "learning_rate": 0.0006, + "loss": 4.511425971984863, + "step": 3992 + }, + { + "epoch": 55.46113537117904, + "grad_norm": 0.01551489345729351, + "learning_rate": 0.0006, + "loss": 4.551141738891602, + "step": 3993 + }, + { + "epoch": 55.47510917030568, + "grad_norm": 0.015072065405547619, + "learning_rate": 0.0006, + "loss": 4.651115894317627, + "step": 3994 + }, + { + "epoch": 55.48908296943232, + "grad_norm": 0.017010986804962158, + "learning_rate": 0.0006, + "loss": 4.583282470703125, + "step": 3995 + }, + { + "epoch": 55.50305676855895, + "grad_norm": 0.01764831133186817, + "learning_rate": 0.0006, + "loss": 4.593364238739014, + "step": 3996 + }, + { + "epoch": 55.51703056768559, + "grad_norm": 0.01762884296476841, + "learning_rate": 0.0006, + "loss": 4.45671272277832, + "step": 3997 + }, + { + "epoch": 55.531004366812226, + "grad_norm": 0.01600457727909088, + "learning_rate": 0.0006, + "loss": 4.511260032653809, + "step": 3998 + }, + { + "epoch": 55.544978165938865, + "grad_norm": 0.015706021338701248, + "learning_rate": 0.0006, + "loss": 4.47593355178833, + "step": 3999 + }, + { + "epoch": 55.5589519650655, + "grad_norm": 0.017718996852636337, + "learning_rate": 0.0006, + "loss": 4.610543251037598, + "step": 4000 + }, + { + "epoch": 55.57292576419214, + "grad_norm": 0.015073757618665695, + "learning_rate": 0.0006, + "loss": 4.466620445251465, + "step": 4001 + }, + { + "epoch": 55.58689956331878, + "grad_norm": 0.01789247989654541, + "learning_rate": 0.0006, + "loss": 4.503409385681152, + "step": 4002 + }, + { + "epoch": 55.60087336244541, + "grad_norm": 0.022329598665237427, + "learning_rate": 0.0006, + "loss": 4.62099552154541, + "step": 4003 + }, + { + "epoch": 55.61484716157205, + "grad_norm": 0.02343829534947872, + "learning_rate": 0.0006, + "loss": 4.620813369750977, + "step": 4004 + }, + { + "epoch": 55.62882096069869, + "grad_norm": 0.023279687389731407, + "learning_rate": 0.0006, + "loss": 4.545765399932861, + "step": 4005 + }, + { + "epoch": 55.64279475982533, + "grad_norm": 0.020817464217543602, + "learning_rate": 0.0006, + "loss": 4.561470031738281, + "step": 4006 + }, + { + "epoch": 55.65676855895197, + "grad_norm": 0.020270058885216713, + "learning_rate": 0.0006, + "loss": 4.5875020027160645, + "step": 4007 + }, + { + "epoch": 55.670742358078606, + "grad_norm": 0.021085843443870544, + "learning_rate": 0.0006, + "loss": 4.504560470581055, + "step": 4008 + }, + { + "epoch": 55.68471615720524, + "grad_norm": 0.018507730215787888, + "learning_rate": 0.0006, + "loss": 4.520644664764404, + "step": 4009 + }, + { + "epoch": 55.698689956331876, + "grad_norm": 0.020571516826748848, + "learning_rate": 0.0006, + "loss": 4.539643287658691, + "step": 4010 + }, + { + "epoch": 55.712663755458514, + "grad_norm": 0.021484510973095894, + "learning_rate": 0.0006, + "loss": 4.628833293914795, + "step": 4011 + }, + { + "epoch": 55.72663755458515, + "grad_norm": 0.017693819478154182, + "learning_rate": 0.0006, + "loss": 4.6023783683776855, + "step": 4012 + }, + { + "epoch": 55.74061135371179, + "grad_norm": 0.020186059176921844, + "learning_rate": 0.0006, + "loss": 4.634884834289551, + "step": 4013 + }, + { + "epoch": 55.75458515283843, + "grad_norm": 0.01835780404508114, + "learning_rate": 0.0006, + "loss": 4.577791213989258, + "step": 4014 + }, + { + "epoch": 55.76855895196506, + "grad_norm": 0.018162988126277924, + "learning_rate": 0.0006, + "loss": 4.641597270965576, + "step": 4015 + }, + { + "epoch": 55.7825327510917, + "grad_norm": 0.018909504637122154, + "learning_rate": 0.0006, + "loss": 4.487447738647461, + "step": 4016 + }, + { + "epoch": 55.79650655021834, + "grad_norm": 0.018972251564264297, + "learning_rate": 0.0006, + "loss": 4.590317726135254, + "step": 4017 + }, + { + "epoch": 55.81048034934498, + "grad_norm": 0.01793607883155346, + "learning_rate": 0.0006, + "loss": 4.4994611740112305, + "step": 4018 + }, + { + "epoch": 55.82445414847162, + "grad_norm": 0.02123514749109745, + "learning_rate": 0.0006, + "loss": 4.479006767272949, + "step": 4019 + }, + { + "epoch": 55.838427947598255, + "grad_norm": 0.021246599033474922, + "learning_rate": 0.0006, + "loss": 4.4738640785217285, + "step": 4020 + }, + { + "epoch": 55.852401746724894, + "grad_norm": 0.020131602883338928, + "learning_rate": 0.0006, + "loss": 4.600308418273926, + "step": 4021 + }, + { + "epoch": 55.866375545851525, + "grad_norm": 0.019159460440278053, + "learning_rate": 0.0006, + "loss": 4.540908336639404, + "step": 4022 + }, + { + "epoch": 55.880349344978164, + "grad_norm": 0.023114705458283424, + "learning_rate": 0.0006, + "loss": 4.641247749328613, + "step": 4023 + }, + { + "epoch": 55.8943231441048, + "grad_norm": 0.02296135015785694, + "learning_rate": 0.0006, + "loss": 4.444624423980713, + "step": 4024 + }, + { + "epoch": 55.90829694323144, + "grad_norm": 0.02204655110836029, + "learning_rate": 0.0006, + "loss": 4.506349563598633, + "step": 4025 + }, + { + "epoch": 55.92227074235808, + "grad_norm": 0.019338225945830345, + "learning_rate": 0.0006, + "loss": 4.616047382354736, + "step": 4026 + }, + { + "epoch": 55.93624454148472, + "grad_norm": 0.020357884466648102, + "learning_rate": 0.0006, + "loss": 4.48170280456543, + "step": 4027 + }, + { + "epoch": 55.95021834061135, + "grad_norm": 0.02426772564649582, + "learning_rate": 0.0006, + "loss": 4.576877117156982, + "step": 4028 + }, + { + "epoch": 55.96419213973799, + "grad_norm": 0.023351816460490227, + "learning_rate": 0.0006, + "loss": 4.458885192871094, + "step": 4029 + }, + { + "epoch": 55.97816593886463, + "grad_norm": 0.019892243668437004, + "learning_rate": 0.0006, + "loss": 4.520229816436768, + "step": 4030 + }, + { + "epoch": 55.992139737991266, + "grad_norm": 0.018331089988350868, + "learning_rate": 0.0006, + "loss": 4.432201862335205, + "step": 4031 + }, + { + "epoch": 56.0, + "grad_norm": 0.021300828084349632, + "learning_rate": 0.0006, + "loss": 4.617072582244873, + "step": 4032 + }, + { + "epoch": 56.0, + "eval_loss": 4.811229705810547, + "eval_runtime": 56.4846, + "eval_samples_per_second": 43.233, + "eval_steps_per_second": 1.363, + "step": 4032 + }, + { + "epoch": 56.01397379912664, + "grad_norm": 0.02199401892721653, + "learning_rate": 0.0006, + "loss": 4.513965606689453, + "step": 4033 + }, + { + "epoch": 56.02794759825328, + "grad_norm": 0.022109858691692352, + "learning_rate": 0.0006, + "loss": 4.557568550109863, + "step": 4034 + }, + { + "epoch": 56.041921397379916, + "grad_norm": 0.020103694871068, + "learning_rate": 0.0006, + "loss": 4.569512844085693, + "step": 4035 + }, + { + "epoch": 56.05589519650655, + "grad_norm": 0.01898941583931446, + "learning_rate": 0.0006, + "loss": 4.623138904571533, + "step": 4036 + }, + { + "epoch": 56.069868995633186, + "grad_norm": 0.02227010577917099, + "learning_rate": 0.0006, + "loss": 4.433316230773926, + "step": 4037 + }, + { + "epoch": 56.083842794759825, + "grad_norm": 0.024683045223355293, + "learning_rate": 0.0006, + "loss": 4.458296775817871, + "step": 4038 + }, + { + "epoch": 56.09781659388646, + "grad_norm": 0.02900184690952301, + "learning_rate": 0.0006, + "loss": 4.531060218811035, + "step": 4039 + }, + { + "epoch": 56.1117903930131, + "grad_norm": 0.025557566434144974, + "learning_rate": 0.0006, + "loss": 4.56685733795166, + "step": 4040 + }, + { + "epoch": 56.12576419213974, + "grad_norm": 0.021414656192064285, + "learning_rate": 0.0006, + "loss": 4.469037055969238, + "step": 4041 + }, + { + "epoch": 56.13973799126637, + "grad_norm": 0.023262616246938705, + "learning_rate": 0.0006, + "loss": 4.493876934051514, + "step": 4042 + }, + { + "epoch": 56.15371179039301, + "grad_norm": 0.023011326789855957, + "learning_rate": 0.0006, + "loss": 4.4572906494140625, + "step": 4043 + }, + { + "epoch": 56.16768558951965, + "grad_norm": 0.02177615463733673, + "learning_rate": 0.0006, + "loss": 4.615389823913574, + "step": 4044 + }, + { + "epoch": 56.18165938864629, + "grad_norm": 0.021527713164687157, + "learning_rate": 0.0006, + "loss": 4.613762855529785, + "step": 4045 + }, + { + "epoch": 56.19563318777293, + "grad_norm": 0.01982325315475464, + "learning_rate": 0.0006, + "loss": 4.53670597076416, + "step": 4046 + }, + { + "epoch": 56.209606986899566, + "grad_norm": 0.022484654560685158, + "learning_rate": 0.0006, + "loss": 4.605565071105957, + "step": 4047 + }, + { + "epoch": 56.223580786026204, + "grad_norm": 0.02346140146255493, + "learning_rate": 0.0006, + "loss": 4.6265950202941895, + "step": 4048 + }, + { + "epoch": 56.237554585152836, + "grad_norm": 0.01876714453101158, + "learning_rate": 0.0006, + "loss": 4.481525421142578, + "step": 4049 + }, + { + "epoch": 56.251528384279474, + "grad_norm": 0.01926851086318493, + "learning_rate": 0.0006, + "loss": 4.630258083343506, + "step": 4050 + }, + { + "epoch": 56.26550218340611, + "grad_norm": 0.020858589559793472, + "learning_rate": 0.0006, + "loss": 4.5265913009643555, + "step": 4051 + }, + { + "epoch": 56.27947598253275, + "grad_norm": 0.01889280416071415, + "learning_rate": 0.0006, + "loss": 4.581275463104248, + "step": 4052 + }, + { + "epoch": 56.29344978165939, + "grad_norm": 0.01801002025604248, + "learning_rate": 0.0006, + "loss": 4.498557090759277, + "step": 4053 + }, + { + "epoch": 56.30742358078603, + "grad_norm": 0.02252180315554142, + "learning_rate": 0.0006, + "loss": 4.565267086029053, + "step": 4054 + }, + { + "epoch": 56.32139737991266, + "grad_norm": 0.02306070365011692, + "learning_rate": 0.0006, + "loss": 4.3591389656066895, + "step": 4055 + }, + { + "epoch": 56.3353711790393, + "grad_norm": 0.020292513072490692, + "learning_rate": 0.0006, + "loss": 4.407517433166504, + "step": 4056 + }, + { + "epoch": 56.34934497816594, + "grad_norm": 0.019154123961925507, + "learning_rate": 0.0006, + "loss": 4.482261657714844, + "step": 4057 + }, + { + "epoch": 56.36331877729258, + "grad_norm": 0.017338937148451805, + "learning_rate": 0.0006, + "loss": 4.4463582038879395, + "step": 4058 + }, + { + "epoch": 56.377292576419215, + "grad_norm": 0.01730559952557087, + "learning_rate": 0.0006, + "loss": 4.542474746704102, + "step": 4059 + }, + { + "epoch": 56.391266375545854, + "grad_norm": 0.017029426991939545, + "learning_rate": 0.0006, + "loss": 4.552488327026367, + "step": 4060 + }, + { + "epoch": 56.40524017467249, + "grad_norm": 0.01745498552918434, + "learning_rate": 0.0006, + "loss": 4.52928352355957, + "step": 4061 + }, + { + "epoch": 56.419213973799124, + "grad_norm": 0.0168515145778656, + "learning_rate": 0.0006, + "loss": 4.4516801834106445, + "step": 4062 + }, + { + "epoch": 56.43318777292576, + "grad_norm": 0.018545417115092278, + "learning_rate": 0.0006, + "loss": 4.569811820983887, + "step": 4063 + }, + { + "epoch": 56.4471615720524, + "grad_norm": 0.02227838523685932, + "learning_rate": 0.0006, + "loss": 4.412224769592285, + "step": 4064 + }, + { + "epoch": 56.46113537117904, + "grad_norm": 0.023338504135608673, + "learning_rate": 0.0006, + "loss": 4.490184307098389, + "step": 4065 + }, + { + "epoch": 56.47510917030568, + "grad_norm": 0.02257535606622696, + "learning_rate": 0.0006, + "loss": 4.479152679443359, + "step": 4066 + }, + { + "epoch": 56.48908296943232, + "grad_norm": 0.022737598046660423, + "learning_rate": 0.0006, + "loss": 4.418235778808594, + "step": 4067 + }, + { + "epoch": 56.50305676855895, + "grad_norm": 0.024603573605418205, + "learning_rate": 0.0006, + "loss": 4.45781135559082, + "step": 4068 + }, + { + "epoch": 56.51703056768559, + "grad_norm": 0.024953778833150864, + "learning_rate": 0.0006, + "loss": 4.593907833099365, + "step": 4069 + }, + { + "epoch": 56.531004366812226, + "grad_norm": 0.02195645309984684, + "learning_rate": 0.0006, + "loss": 4.467643737792969, + "step": 4070 + }, + { + "epoch": 56.544978165938865, + "grad_norm": 0.017126578837633133, + "learning_rate": 0.0006, + "loss": 4.555484771728516, + "step": 4071 + }, + { + "epoch": 56.5589519650655, + "grad_norm": 0.019482605159282684, + "learning_rate": 0.0006, + "loss": 4.430915832519531, + "step": 4072 + }, + { + "epoch": 56.57292576419214, + "grad_norm": 0.0219392292201519, + "learning_rate": 0.0006, + "loss": 4.593365669250488, + "step": 4073 + }, + { + "epoch": 56.58689956331878, + "grad_norm": 0.02420145645737648, + "learning_rate": 0.0006, + "loss": 4.491467475891113, + "step": 4074 + }, + { + "epoch": 56.60087336244541, + "grad_norm": 0.021523723378777504, + "learning_rate": 0.0006, + "loss": 4.463634490966797, + "step": 4075 + }, + { + "epoch": 56.61484716157205, + "grad_norm": 0.020242048427462578, + "learning_rate": 0.0006, + "loss": 4.577968597412109, + "step": 4076 + }, + { + "epoch": 56.62882096069869, + "grad_norm": 0.017771173268556595, + "learning_rate": 0.0006, + "loss": 4.568269729614258, + "step": 4077 + }, + { + "epoch": 56.64279475982533, + "grad_norm": 0.01598544418811798, + "learning_rate": 0.0006, + "loss": 4.476841926574707, + "step": 4078 + }, + { + "epoch": 56.65676855895197, + "grad_norm": 0.0168446097522974, + "learning_rate": 0.0006, + "loss": 4.480965614318848, + "step": 4079 + }, + { + "epoch": 56.670742358078606, + "grad_norm": 0.014948660507798195, + "learning_rate": 0.0006, + "loss": 4.535400390625, + "step": 4080 + }, + { + "epoch": 56.68471615720524, + "grad_norm": 0.014899153262376785, + "learning_rate": 0.0006, + "loss": 4.476048946380615, + "step": 4081 + }, + { + "epoch": 56.698689956331876, + "grad_norm": 0.014051870442926884, + "learning_rate": 0.0006, + "loss": 4.533884048461914, + "step": 4082 + }, + { + "epoch": 56.712663755458514, + "grad_norm": 0.014708572067320347, + "learning_rate": 0.0006, + "loss": 4.586690902709961, + "step": 4083 + }, + { + "epoch": 56.72663755458515, + "grad_norm": 0.014797299169003963, + "learning_rate": 0.0006, + "loss": 4.446291446685791, + "step": 4084 + }, + { + "epoch": 56.74061135371179, + "grad_norm": 0.015256541781127453, + "learning_rate": 0.0006, + "loss": 4.553572654724121, + "step": 4085 + }, + { + "epoch": 56.75458515283843, + "grad_norm": 0.016042491421103477, + "learning_rate": 0.0006, + "loss": 4.507801055908203, + "step": 4086 + }, + { + "epoch": 56.76855895196506, + "grad_norm": 0.018835173919796944, + "learning_rate": 0.0006, + "loss": 4.393911838531494, + "step": 4087 + }, + { + "epoch": 56.7825327510917, + "grad_norm": 0.021994033828377724, + "learning_rate": 0.0006, + "loss": 4.543368339538574, + "step": 4088 + }, + { + "epoch": 56.79650655021834, + "grad_norm": 0.0230990182608366, + "learning_rate": 0.0006, + "loss": 4.507420539855957, + "step": 4089 + }, + { + "epoch": 56.81048034934498, + "grad_norm": 0.018818842247128487, + "learning_rate": 0.0006, + "loss": 4.499941349029541, + "step": 4090 + }, + { + "epoch": 56.82445414847162, + "grad_norm": 0.018257278949022293, + "learning_rate": 0.0006, + "loss": 4.380057334899902, + "step": 4091 + }, + { + "epoch": 56.838427947598255, + "grad_norm": 0.020073018968105316, + "learning_rate": 0.0006, + "loss": 4.572811603546143, + "step": 4092 + }, + { + "epoch": 56.852401746724894, + "grad_norm": 0.02078658528625965, + "learning_rate": 0.0006, + "loss": 4.446835517883301, + "step": 4093 + }, + { + "epoch": 56.866375545851525, + "grad_norm": 0.01793118566274643, + "learning_rate": 0.0006, + "loss": 4.517854690551758, + "step": 4094 + }, + { + "epoch": 56.880349344978164, + "grad_norm": 0.01608281210064888, + "learning_rate": 0.0006, + "loss": 4.506482124328613, + "step": 4095 + }, + { + "epoch": 56.8943231441048, + "grad_norm": 0.017405368387699127, + "learning_rate": 0.0006, + "loss": 4.439718246459961, + "step": 4096 + }, + { + "epoch": 56.90829694323144, + "grad_norm": 0.017930863425135612, + "learning_rate": 0.0006, + "loss": 4.409365653991699, + "step": 4097 + }, + { + "epoch": 56.92227074235808, + "grad_norm": 0.015757670626044273, + "learning_rate": 0.0006, + "loss": 4.505547523498535, + "step": 4098 + }, + { + "epoch": 56.93624454148472, + "grad_norm": 0.01580345816910267, + "learning_rate": 0.0006, + "loss": 4.462774753570557, + "step": 4099 + }, + { + "epoch": 56.95021834061135, + "grad_norm": 0.01659180410206318, + "learning_rate": 0.0006, + "loss": 4.49345588684082, + "step": 4100 + }, + { + "epoch": 56.96419213973799, + "grad_norm": 0.01877259835600853, + "learning_rate": 0.0006, + "loss": 4.574443340301514, + "step": 4101 + }, + { + "epoch": 56.97816593886463, + "grad_norm": 0.01977839693427086, + "learning_rate": 0.0006, + "loss": 4.574034690856934, + "step": 4102 + }, + { + "epoch": 56.992139737991266, + "grad_norm": 0.020627597346901894, + "learning_rate": 0.0006, + "loss": 4.50969123840332, + "step": 4103 + }, + { + "epoch": 57.0, + "grad_norm": 0.025128807872533798, + "learning_rate": 0.0006, + "loss": 4.401662826538086, + "step": 4104 + }, + { + "epoch": 57.0, + "eval_loss": 4.779252052307129, + "eval_runtime": 56.8104, + "eval_samples_per_second": 42.985, + "eval_steps_per_second": 1.355, + "step": 4104 + }, + { + "epoch": 57.01397379912664, + "grad_norm": 0.027712570503354073, + "learning_rate": 0.0006, + "loss": 4.472955703735352, + "step": 4105 + }, + { + "epoch": 57.02794759825328, + "grad_norm": 0.02164197526872158, + "learning_rate": 0.0006, + "loss": 4.387596130371094, + "step": 4106 + }, + { + "epoch": 57.041921397379916, + "grad_norm": 0.017972752451896667, + "learning_rate": 0.0006, + "loss": 4.4326276779174805, + "step": 4107 + }, + { + "epoch": 57.05589519650655, + "grad_norm": 0.020811019465327263, + "learning_rate": 0.0006, + "loss": 4.452951431274414, + "step": 4108 + }, + { + "epoch": 57.069868995633186, + "grad_norm": 0.019913259893655777, + "learning_rate": 0.0006, + "loss": 4.448301315307617, + "step": 4109 + }, + { + "epoch": 57.083842794759825, + "grad_norm": 0.017409102991223335, + "learning_rate": 0.0006, + "loss": 4.4108357429504395, + "step": 4110 + }, + { + "epoch": 57.09781659388646, + "grad_norm": 0.019374269992113113, + "learning_rate": 0.0006, + "loss": 4.398412227630615, + "step": 4111 + }, + { + "epoch": 57.1117903930131, + "grad_norm": 0.015521776862442493, + "learning_rate": 0.0006, + "loss": 4.508594989776611, + "step": 4112 + }, + { + "epoch": 57.12576419213974, + "grad_norm": 0.017201920971274376, + "learning_rate": 0.0006, + "loss": 4.449558734893799, + "step": 4113 + }, + { + "epoch": 57.13973799126637, + "grad_norm": 0.018062541261315346, + "learning_rate": 0.0006, + "loss": 4.517874240875244, + "step": 4114 + }, + { + "epoch": 57.15371179039301, + "grad_norm": 0.015846073627471924, + "learning_rate": 0.0006, + "loss": 4.427501678466797, + "step": 4115 + }, + { + "epoch": 57.16768558951965, + "grad_norm": 0.01849362626671791, + "learning_rate": 0.0006, + "loss": 4.433181285858154, + "step": 4116 + }, + { + "epoch": 57.18165938864629, + "grad_norm": 0.01948804222047329, + "learning_rate": 0.0006, + "loss": 4.446122646331787, + "step": 4117 + }, + { + "epoch": 57.19563318777293, + "grad_norm": 0.017354389652609825, + "learning_rate": 0.0006, + "loss": 4.431153774261475, + "step": 4118 + }, + { + "epoch": 57.209606986899566, + "grad_norm": 0.015817373991012573, + "learning_rate": 0.0006, + "loss": 4.448566913604736, + "step": 4119 + }, + { + "epoch": 57.223580786026204, + "grad_norm": 0.01635785959661007, + "learning_rate": 0.0006, + "loss": 4.408031463623047, + "step": 4120 + }, + { + "epoch": 57.237554585152836, + "grad_norm": 0.014528338797390461, + "learning_rate": 0.0006, + "loss": 4.541668891906738, + "step": 4121 + }, + { + "epoch": 57.251528384279474, + "grad_norm": 0.01589174196124077, + "learning_rate": 0.0006, + "loss": 4.513199329376221, + "step": 4122 + }, + { + "epoch": 57.26550218340611, + "grad_norm": 0.016093695536255836, + "learning_rate": 0.0006, + "loss": 4.516178131103516, + "step": 4123 + }, + { + "epoch": 57.27947598253275, + "grad_norm": 0.01788361556828022, + "learning_rate": 0.0006, + "loss": 4.524880409240723, + "step": 4124 + }, + { + "epoch": 57.29344978165939, + "grad_norm": 0.01591576263308525, + "learning_rate": 0.0006, + "loss": 4.44764518737793, + "step": 4125 + }, + { + "epoch": 57.30742358078603, + "grad_norm": 0.015452570281922817, + "learning_rate": 0.0006, + "loss": 4.46844482421875, + "step": 4126 + }, + { + "epoch": 57.32139737991266, + "grad_norm": 0.016404492780566216, + "learning_rate": 0.0006, + "loss": 4.443678855895996, + "step": 4127 + }, + { + "epoch": 57.3353711790393, + "grad_norm": 0.019967148080468178, + "learning_rate": 0.0006, + "loss": 4.430230140686035, + "step": 4128 + }, + { + "epoch": 57.34934497816594, + "grad_norm": 0.02594015561044216, + "learning_rate": 0.0006, + "loss": 4.47061824798584, + "step": 4129 + }, + { + "epoch": 57.36331877729258, + "grad_norm": 0.031096890568733215, + "learning_rate": 0.0006, + "loss": 4.462275981903076, + "step": 4130 + }, + { + "epoch": 57.377292576419215, + "grad_norm": 0.02834898792207241, + "learning_rate": 0.0006, + "loss": 4.41143798828125, + "step": 4131 + }, + { + "epoch": 57.391266375545854, + "grad_norm": 0.023621153086423874, + "learning_rate": 0.0006, + "loss": 4.562548637390137, + "step": 4132 + }, + { + "epoch": 57.40524017467249, + "grad_norm": 0.02277296595275402, + "learning_rate": 0.0006, + "loss": 4.456247329711914, + "step": 4133 + }, + { + "epoch": 57.419213973799124, + "grad_norm": 0.019868768751621246, + "learning_rate": 0.0006, + "loss": 4.524300575256348, + "step": 4134 + }, + { + "epoch": 57.43318777292576, + "grad_norm": 0.01887078955769539, + "learning_rate": 0.0006, + "loss": 4.540097236633301, + "step": 4135 + }, + { + "epoch": 57.4471615720524, + "grad_norm": 0.024268802255392075, + "learning_rate": 0.0006, + "loss": 4.397233963012695, + "step": 4136 + }, + { + "epoch": 57.46113537117904, + "grad_norm": 0.025739721953868866, + "learning_rate": 0.0006, + "loss": 4.354194641113281, + "step": 4137 + }, + { + "epoch": 57.47510917030568, + "grad_norm": 0.02442290261387825, + "learning_rate": 0.0006, + "loss": 4.303822994232178, + "step": 4138 + }, + { + "epoch": 57.48908296943232, + "grad_norm": 0.022215209901332855, + "learning_rate": 0.0006, + "loss": 4.4870452880859375, + "step": 4139 + }, + { + "epoch": 57.50305676855895, + "grad_norm": 0.01789158768951893, + "learning_rate": 0.0006, + "loss": 4.343809127807617, + "step": 4140 + }, + { + "epoch": 57.51703056768559, + "grad_norm": 0.020381739363074303, + "learning_rate": 0.0006, + "loss": 4.53312873840332, + "step": 4141 + }, + { + "epoch": 57.531004366812226, + "grad_norm": 0.019585080444812775, + "learning_rate": 0.0006, + "loss": 4.510162353515625, + "step": 4142 + }, + { + "epoch": 57.544978165938865, + "grad_norm": 0.01985127478837967, + "learning_rate": 0.0006, + "loss": 4.3654022216796875, + "step": 4143 + }, + { + "epoch": 57.5589519650655, + "grad_norm": 0.01962905190885067, + "learning_rate": 0.0006, + "loss": 4.560830116271973, + "step": 4144 + }, + { + "epoch": 57.57292576419214, + "grad_norm": 0.02679629437625408, + "learning_rate": 0.0006, + "loss": 4.300804138183594, + "step": 4145 + }, + { + "epoch": 57.58689956331878, + "grad_norm": 0.039830856025218964, + "learning_rate": 0.0006, + "loss": 4.54150915145874, + "step": 4146 + }, + { + "epoch": 57.60087336244541, + "grad_norm": 0.049527477473020554, + "learning_rate": 0.0006, + "loss": 4.405792236328125, + "step": 4147 + }, + { + "epoch": 57.61484716157205, + "grad_norm": 0.042966946959495544, + "learning_rate": 0.0006, + "loss": 4.500643730163574, + "step": 4148 + }, + { + "epoch": 57.62882096069869, + "grad_norm": 0.08999177813529968, + "learning_rate": 0.0006, + "loss": 4.504790782928467, + "step": 4149 + }, + { + "epoch": 57.64279475982533, + "grad_norm": 0.44693124294281006, + "learning_rate": 0.0006, + "loss": 4.845344066619873, + "step": 4150 + }, + { + "epoch": 57.65676855895197, + "grad_norm": 0.7143144011497498, + "learning_rate": 0.0006, + "loss": 6.907845497131348, + "step": 4151 + }, + { + "epoch": 57.670742358078606, + "grad_norm": 0.25239595770835876, + "learning_rate": 0.0006, + "loss": 7.082579612731934, + "step": 4152 + }, + { + "epoch": 57.68471615720524, + "grad_norm": 0.26550790667533875, + "learning_rate": 0.0006, + "loss": 7.50180196762085, + "step": 4153 + }, + { + "epoch": 57.698689956331876, + "grad_norm": 0.14735378324985504, + "learning_rate": 0.0006, + "loss": 7.214590072631836, + "step": 4154 + }, + { + "epoch": 57.712663755458514, + "grad_norm": 0.17685334384441376, + "learning_rate": 0.0006, + "loss": 6.861320495605469, + "step": 4155 + }, + { + "epoch": 57.72663755458515, + "grad_norm": 0.18231339752674103, + "learning_rate": 0.0006, + "loss": 6.791839599609375, + "step": 4156 + }, + { + "epoch": 57.74061135371179, + "grad_norm": 0.09230359643697739, + "learning_rate": 0.0006, + "loss": 6.760223388671875, + "step": 4157 + }, + { + "epoch": 57.75458515283843, + "grad_norm": 0.06918249279260635, + "learning_rate": 0.0006, + "loss": 6.532999515533447, + "step": 4158 + }, + { + "epoch": 57.76855895196506, + "grad_norm": 0.0901007279753685, + "learning_rate": 0.0006, + "loss": 6.52907133102417, + "step": 4159 + }, + { + "epoch": 57.7825327510917, + "grad_norm": 0.05678229779005051, + "learning_rate": 0.0006, + "loss": 6.295290470123291, + "step": 4160 + }, + { + "epoch": 57.79650655021834, + "grad_norm": 0.04450210556387901, + "learning_rate": 0.0006, + "loss": 6.14900541305542, + "step": 4161 + }, + { + "epoch": 57.81048034934498, + "grad_norm": 0.049191512167453766, + "learning_rate": 0.0006, + "loss": 6.19650411605835, + "step": 4162 + }, + { + "epoch": 57.82445414847162, + "grad_norm": 0.03908967226743698, + "learning_rate": 0.0006, + "loss": 6.054948806762695, + "step": 4163 + }, + { + "epoch": 57.838427947598255, + "grad_norm": 0.04023387283086777, + "learning_rate": 0.0006, + "loss": 6.034873962402344, + "step": 4164 + }, + { + "epoch": 57.852401746724894, + "grad_norm": 0.03304268419742584, + "learning_rate": 0.0006, + "loss": 5.769852161407471, + "step": 4165 + }, + { + "epoch": 57.866375545851525, + "grad_norm": 0.03432450816035271, + "learning_rate": 0.0006, + "loss": 5.788785457611084, + "step": 4166 + }, + { + "epoch": 57.880349344978164, + "grad_norm": 0.029965840280056, + "learning_rate": 0.0006, + "loss": 5.815243721008301, + "step": 4167 + }, + { + "epoch": 57.8943231441048, + "grad_norm": 0.03637698292732239, + "learning_rate": 0.0006, + "loss": 5.7204060554504395, + "step": 4168 + }, + { + "epoch": 57.90829694323144, + "grad_norm": 0.03477517142891884, + "learning_rate": 0.0006, + "loss": 5.605930328369141, + "step": 4169 + }, + { + "epoch": 57.92227074235808, + "grad_norm": 0.05061859264969826, + "learning_rate": 0.0006, + "loss": 5.547054767608643, + "step": 4170 + }, + { + "epoch": 57.93624454148472, + "grad_norm": 0.06299655884504318, + "learning_rate": 0.0006, + "loss": 5.552166938781738, + "step": 4171 + }, + { + "epoch": 57.95021834061135, + "grad_norm": 0.04215948283672333, + "learning_rate": 0.0006, + "loss": 5.375457763671875, + "step": 4172 + }, + { + "epoch": 57.96419213973799, + "grad_norm": 0.0348566472530365, + "learning_rate": 0.0006, + "loss": 5.418606281280518, + "step": 4173 + }, + { + "epoch": 57.97816593886463, + "grad_norm": 0.03807530924677849, + "learning_rate": 0.0006, + "loss": 5.403156280517578, + "step": 4174 + }, + { + "epoch": 57.992139737991266, + "grad_norm": 0.02989993803203106, + "learning_rate": 0.0006, + "loss": 5.355951309204102, + "step": 4175 + }, + { + "epoch": 58.0, + "grad_norm": 0.034300558269023895, + "learning_rate": 0.0006, + "loss": 5.44920539855957, + "step": 4176 + }, + { + "epoch": 58.0, + "eval_loss": 5.416731834411621, + "eval_runtime": 56.7934, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4176 + }, + { + "epoch": 58.01397379912664, + "grad_norm": 0.033961448818445206, + "learning_rate": 0.0006, + "loss": 5.250707626342773, + "step": 4177 + }, + { + "epoch": 58.02794759825328, + "grad_norm": 0.027343502268195152, + "learning_rate": 0.0006, + "loss": 5.149238109588623, + "step": 4178 + }, + { + "epoch": 58.041921397379916, + "grad_norm": 0.03033650480210781, + "learning_rate": 0.0006, + "loss": 5.072463512420654, + "step": 4179 + }, + { + "epoch": 58.05589519650655, + "grad_norm": 0.028262868523597717, + "learning_rate": 0.0006, + "loss": 5.109417915344238, + "step": 4180 + }, + { + "epoch": 58.069868995633186, + "grad_norm": 0.02373034693300724, + "learning_rate": 0.0006, + "loss": 5.082464218139648, + "step": 4181 + }, + { + "epoch": 58.083842794759825, + "grad_norm": 0.025879688560962677, + "learning_rate": 0.0006, + "loss": 5.056792259216309, + "step": 4182 + }, + { + "epoch": 58.09781659388646, + "grad_norm": 0.026252275332808495, + "learning_rate": 0.0006, + "loss": 5.064239978790283, + "step": 4183 + }, + { + "epoch": 58.1117903930131, + "grad_norm": 0.028177792206406593, + "learning_rate": 0.0006, + "loss": 4.922310829162598, + "step": 4184 + }, + { + "epoch": 58.12576419213974, + "grad_norm": 0.02630820870399475, + "learning_rate": 0.0006, + "loss": 4.992884635925293, + "step": 4185 + }, + { + "epoch": 58.13973799126637, + "grad_norm": 0.024890903383493423, + "learning_rate": 0.0006, + "loss": 4.910036563873291, + "step": 4186 + }, + { + "epoch": 58.15371179039301, + "grad_norm": 0.02822992391884327, + "learning_rate": 0.0006, + "loss": 4.866640090942383, + "step": 4187 + }, + { + "epoch": 58.16768558951965, + "grad_norm": 0.03253607079386711, + "learning_rate": 0.0006, + "loss": 4.9112958908081055, + "step": 4188 + }, + { + "epoch": 58.18165938864629, + "grad_norm": 0.03765944764018059, + "learning_rate": 0.0006, + "loss": 4.774468421936035, + "step": 4189 + }, + { + "epoch": 58.19563318777293, + "grad_norm": 0.039874881505966187, + "learning_rate": 0.0006, + "loss": 4.831678867340088, + "step": 4190 + }, + { + "epoch": 58.209606986899566, + "grad_norm": 0.034595925360918045, + "learning_rate": 0.0006, + "loss": 4.8243794441223145, + "step": 4191 + }, + { + "epoch": 58.223580786026204, + "grad_norm": 0.04840189591050148, + "learning_rate": 0.0006, + "loss": 4.854156494140625, + "step": 4192 + }, + { + "epoch": 58.237554585152836, + "grad_norm": 0.06358753889799118, + "learning_rate": 0.0006, + "loss": 4.836430549621582, + "step": 4193 + }, + { + "epoch": 58.251528384279474, + "grad_norm": 0.0507555715739727, + "learning_rate": 0.0006, + "loss": 4.88357400894165, + "step": 4194 + }, + { + "epoch": 58.26550218340611, + "grad_norm": 0.04384690150618553, + "learning_rate": 0.0006, + "loss": 4.774049282073975, + "step": 4195 + }, + { + "epoch": 58.27947598253275, + "grad_norm": 0.03246928006410599, + "learning_rate": 0.0006, + "loss": 4.686605930328369, + "step": 4196 + }, + { + "epoch": 58.29344978165939, + "grad_norm": 0.028878789395093918, + "learning_rate": 0.0006, + "loss": 4.796070575714111, + "step": 4197 + }, + { + "epoch": 58.30742358078603, + "grad_norm": 0.028175361454486847, + "learning_rate": 0.0006, + "loss": 4.728490829467773, + "step": 4198 + }, + { + "epoch": 58.32139737991266, + "grad_norm": 0.029428014531731606, + "learning_rate": 0.0006, + "loss": 4.7159624099731445, + "step": 4199 + }, + { + "epoch": 58.3353711790393, + "grad_norm": 0.031227827072143555, + "learning_rate": 0.0006, + "loss": 4.702625274658203, + "step": 4200 + }, + { + "epoch": 58.34934497816594, + "grad_norm": 0.028843428939580917, + "learning_rate": 0.0006, + "loss": 4.811799049377441, + "step": 4201 + }, + { + "epoch": 58.36331877729258, + "grad_norm": 0.026780391111969948, + "learning_rate": 0.0006, + "loss": 4.81662654876709, + "step": 4202 + }, + { + "epoch": 58.377292576419215, + "grad_norm": 0.019699757918715477, + "learning_rate": 0.0006, + "loss": 4.713685989379883, + "step": 4203 + }, + { + "epoch": 58.391266375545854, + "grad_norm": 0.024389250203967094, + "learning_rate": 0.0006, + "loss": 4.72354793548584, + "step": 4204 + }, + { + "epoch": 58.40524017467249, + "grad_norm": 0.02218274027109146, + "learning_rate": 0.0006, + "loss": 4.674125671386719, + "step": 4205 + }, + { + "epoch": 58.419213973799124, + "grad_norm": 0.02041488140821457, + "learning_rate": 0.0006, + "loss": 4.7909698486328125, + "step": 4206 + }, + { + "epoch": 58.43318777292576, + "grad_norm": 0.018561935052275658, + "learning_rate": 0.0006, + "loss": 4.808770179748535, + "step": 4207 + }, + { + "epoch": 58.4471615720524, + "grad_norm": 0.01708846725523472, + "learning_rate": 0.0006, + "loss": 4.701626777648926, + "step": 4208 + }, + { + "epoch": 58.46113537117904, + "grad_norm": 0.0222734697163105, + "learning_rate": 0.0006, + "loss": 4.505410194396973, + "step": 4209 + }, + { + "epoch": 58.47510917030568, + "grad_norm": 0.03197057917714119, + "learning_rate": 0.0006, + "loss": 4.610360145568848, + "step": 4210 + }, + { + "epoch": 58.48908296943232, + "grad_norm": 0.06854520738124847, + "learning_rate": 0.0006, + "loss": 4.625649452209473, + "step": 4211 + }, + { + "epoch": 58.50305676855895, + "grad_norm": 0.09981521219015121, + "learning_rate": 0.0006, + "loss": 4.832150459289551, + "step": 4212 + }, + { + "epoch": 58.51703056768559, + "grad_norm": 0.046356040984392166, + "learning_rate": 0.0006, + "loss": 4.5641632080078125, + "step": 4213 + }, + { + "epoch": 58.531004366812226, + "grad_norm": 0.03882667422294617, + "learning_rate": 0.0006, + "loss": 4.666974067687988, + "step": 4214 + }, + { + "epoch": 58.544978165938865, + "grad_norm": 0.033203233033418655, + "learning_rate": 0.0006, + "loss": 4.629413604736328, + "step": 4215 + }, + { + "epoch": 58.5589519650655, + "grad_norm": 0.02806735597550869, + "learning_rate": 0.0006, + "loss": 4.763245582580566, + "step": 4216 + }, + { + "epoch": 58.57292576419214, + "grad_norm": 0.026728278025984764, + "learning_rate": 0.0006, + "loss": 4.629597187042236, + "step": 4217 + }, + { + "epoch": 58.58689956331878, + "grad_norm": 0.024380959570407867, + "learning_rate": 0.0006, + "loss": 4.633614540100098, + "step": 4218 + }, + { + "epoch": 58.60087336244541, + "grad_norm": 0.022312544286251068, + "learning_rate": 0.0006, + "loss": 4.671379566192627, + "step": 4219 + }, + { + "epoch": 58.61484716157205, + "grad_norm": 0.022049568593502045, + "learning_rate": 0.0006, + "loss": 4.661557197570801, + "step": 4220 + }, + { + "epoch": 58.62882096069869, + "grad_norm": 0.02119818702340126, + "learning_rate": 0.0006, + "loss": 4.688562393188477, + "step": 4221 + }, + { + "epoch": 58.64279475982533, + "grad_norm": 0.019769301638007164, + "learning_rate": 0.0006, + "loss": 4.607748031616211, + "step": 4222 + }, + { + "epoch": 58.65676855895197, + "grad_norm": 0.02124079130589962, + "learning_rate": 0.0006, + "loss": 4.570640563964844, + "step": 4223 + }, + { + "epoch": 58.670742358078606, + "grad_norm": 0.016344040632247925, + "learning_rate": 0.0006, + "loss": 4.676113128662109, + "step": 4224 + }, + { + "epoch": 58.68471615720524, + "grad_norm": 0.01735256239771843, + "learning_rate": 0.0006, + "loss": 4.62088680267334, + "step": 4225 + }, + { + "epoch": 58.698689956331876, + "grad_norm": 0.019323458895087242, + "learning_rate": 0.0006, + "loss": 4.579405307769775, + "step": 4226 + }, + { + "epoch": 58.712663755458514, + "grad_norm": 0.01760544814169407, + "learning_rate": 0.0006, + "loss": 4.625407695770264, + "step": 4227 + }, + { + "epoch": 58.72663755458515, + "grad_norm": 0.015593250282108784, + "learning_rate": 0.0006, + "loss": 4.6326117515563965, + "step": 4228 + }, + { + "epoch": 58.74061135371179, + "grad_norm": 0.014357523061335087, + "learning_rate": 0.0006, + "loss": 4.6869401931762695, + "step": 4229 + }, + { + "epoch": 58.75458515283843, + "grad_norm": 0.01604357920587063, + "learning_rate": 0.0006, + "loss": 4.690784931182861, + "step": 4230 + }, + { + "epoch": 58.76855895196506, + "grad_norm": 0.01673845760524273, + "learning_rate": 0.0006, + "loss": 4.513410568237305, + "step": 4231 + }, + { + "epoch": 58.7825327510917, + "grad_norm": 0.013657779432833195, + "learning_rate": 0.0006, + "loss": 4.637057781219482, + "step": 4232 + }, + { + "epoch": 58.79650655021834, + "grad_norm": 0.013373114168643951, + "learning_rate": 0.0006, + "loss": 4.719967842102051, + "step": 4233 + }, + { + "epoch": 58.81048034934498, + "grad_norm": 0.016345568001270294, + "learning_rate": 0.0006, + "loss": 4.604109764099121, + "step": 4234 + }, + { + "epoch": 58.82445414847162, + "grad_norm": 0.01573033444583416, + "learning_rate": 0.0006, + "loss": 4.6471428871154785, + "step": 4235 + }, + { + "epoch": 58.838427947598255, + "grad_norm": 0.01517449039965868, + "learning_rate": 0.0006, + "loss": 4.722684860229492, + "step": 4236 + }, + { + "epoch": 58.852401746724894, + "grad_norm": 0.0145102022215724, + "learning_rate": 0.0006, + "loss": 4.662420272827148, + "step": 4237 + }, + { + "epoch": 58.866375545851525, + "grad_norm": 0.01457217987626791, + "learning_rate": 0.0006, + "loss": 4.585169792175293, + "step": 4238 + }, + { + "epoch": 58.880349344978164, + "grad_norm": 0.01406773366034031, + "learning_rate": 0.0006, + "loss": 4.5413665771484375, + "step": 4239 + }, + { + "epoch": 58.8943231441048, + "grad_norm": 0.012653871439397335, + "learning_rate": 0.0006, + "loss": 4.553636074066162, + "step": 4240 + }, + { + "epoch": 58.90829694323144, + "grad_norm": 0.01261200476437807, + "learning_rate": 0.0006, + "loss": 4.534193992614746, + "step": 4241 + }, + { + "epoch": 58.92227074235808, + "grad_norm": 0.012955864891409874, + "learning_rate": 0.0006, + "loss": 4.697248458862305, + "step": 4242 + }, + { + "epoch": 58.93624454148472, + "grad_norm": 0.012134749442338943, + "learning_rate": 0.0006, + "loss": 4.557095527648926, + "step": 4243 + }, + { + "epoch": 58.95021834061135, + "grad_norm": 0.012164677493274212, + "learning_rate": 0.0006, + "loss": 4.623579025268555, + "step": 4244 + }, + { + "epoch": 58.96419213973799, + "grad_norm": 0.013802947476506233, + "learning_rate": 0.0006, + "loss": 4.562838077545166, + "step": 4245 + }, + { + "epoch": 58.97816593886463, + "grad_norm": 0.011420476250350475, + "learning_rate": 0.0006, + "loss": 4.62265682220459, + "step": 4246 + }, + { + "epoch": 58.992139737991266, + "grad_norm": 0.01169576682150364, + "learning_rate": 0.0006, + "loss": 4.727449417114258, + "step": 4247 + }, + { + "epoch": 59.0, + "grad_norm": 0.013577710837125778, + "learning_rate": 0.0006, + "loss": 4.536325454711914, + "step": 4248 + }, + { + "epoch": 59.0, + "eval_loss": 4.86527681350708, + "eval_runtime": 60.4463, + "eval_samples_per_second": 40.399, + "eval_steps_per_second": 1.274, + "step": 4248 + }, + { + "epoch": 59.01397379912664, + "grad_norm": 0.015251655131578445, + "learning_rate": 0.0006, + "loss": 4.634598731994629, + "step": 4249 + }, + { + "epoch": 59.02794759825328, + "grad_norm": 0.017414938658475876, + "learning_rate": 0.0006, + "loss": 4.521329879760742, + "step": 4250 + }, + { + "epoch": 59.041921397379916, + "grad_norm": 0.01656588353216648, + "learning_rate": 0.0006, + "loss": 4.446192264556885, + "step": 4251 + }, + { + "epoch": 59.05589519650655, + "grad_norm": 0.012038925662636757, + "learning_rate": 0.0006, + "loss": 4.523855209350586, + "step": 4252 + }, + { + "epoch": 59.069868995633186, + "grad_norm": 0.014403033070266247, + "learning_rate": 0.0006, + "loss": 4.561079978942871, + "step": 4253 + }, + { + "epoch": 59.083842794759825, + "grad_norm": 0.016243597492575645, + "learning_rate": 0.0006, + "loss": 4.673602104187012, + "step": 4254 + }, + { + "epoch": 59.09781659388646, + "grad_norm": 0.016492877155542374, + "learning_rate": 0.0006, + "loss": 4.564824104309082, + "step": 4255 + }, + { + "epoch": 59.1117903930131, + "grad_norm": 0.015980370342731476, + "learning_rate": 0.0006, + "loss": 4.339249134063721, + "step": 4256 + }, + { + "epoch": 59.12576419213974, + "grad_norm": 0.019687356427311897, + "learning_rate": 0.0006, + "loss": 4.481834888458252, + "step": 4257 + }, + { + "epoch": 59.13973799126637, + "grad_norm": 0.02983798086643219, + "learning_rate": 0.0006, + "loss": 4.4495697021484375, + "step": 4258 + }, + { + "epoch": 59.15371179039301, + "grad_norm": 0.04600981995463371, + "learning_rate": 0.0006, + "loss": 4.567838668823242, + "step": 4259 + }, + { + "epoch": 59.16768558951965, + "grad_norm": 0.04852001741528511, + "learning_rate": 0.0006, + "loss": 4.594701766967773, + "step": 4260 + }, + { + "epoch": 59.18165938864629, + "grad_norm": 0.03429504856467247, + "learning_rate": 0.0006, + "loss": 4.573735237121582, + "step": 4261 + }, + { + "epoch": 59.19563318777293, + "grad_norm": 0.029345160350203514, + "learning_rate": 0.0006, + "loss": 4.488258361816406, + "step": 4262 + }, + { + "epoch": 59.209606986899566, + "grad_norm": 0.01961623504757881, + "learning_rate": 0.0006, + "loss": 4.545222282409668, + "step": 4263 + }, + { + "epoch": 59.223580786026204, + "grad_norm": 0.020406130701303482, + "learning_rate": 0.0006, + "loss": 4.471898078918457, + "step": 4264 + }, + { + "epoch": 59.237554585152836, + "grad_norm": 0.020500272512435913, + "learning_rate": 0.0006, + "loss": 4.538337707519531, + "step": 4265 + }, + { + "epoch": 59.251528384279474, + "grad_norm": 0.020727017894387245, + "learning_rate": 0.0006, + "loss": 4.437284469604492, + "step": 4266 + }, + { + "epoch": 59.26550218340611, + "grad_norm": 0.03395998477935791, + "learning_rate": 0.0006, + "loss": 4.459124565124512, + "step": 4267 + }, + { + "epoch": 59.27947598253275, + "grad_norm": 0.04635605961084366, + "learning_rate": 0.0006, + "loss": 4.520228385925293, + "step": 4268 + }, + { + "epoch": 59.29344978165939, + "grad_norm": 0.04253006726503372, + "learning_rate": 0.0006, + "loss": 4.578673362731934, + "step": 4269 + }, + { + "epoch": 59.30742358078603, + "grad_norm": 0.02403509058058262, + "learning_rate": 0.0006, + "loss": 4.484639644622803, + "step": 4270 + }, + { + "epoch": 59.32139737991266, + "grad_norm": 0.02254190482199192, + "learning_rate": 0.0006, + "loss": 4.511314868927002, + "step": 4271 + }, + { + "epoch": 59.3353711790393, + "grad_norm": 0.021585190668702126, + "learning_rate": 0.0006, + "loss": 4.530239105224609, + "step": 4272 + }, + { + "epoch": 59.34934497816594, + "grad_norm": 0.021210316568613052, + "learning_rate": 0.0006, + "loss": 4.593514442443848, + "step": 4273 + }, + { + "epoch": 59.36331877729258, + "grad_norm": 0.019317107275128365, + "learning_rate": 0.0006, + "loss": 4.5591535568237305, + "step": 4274 + }, + { + "epoch": 59.377292576419215, + "grad_norm": 0.018059708178043365, + "learning_rate": 0.0006, + "loss": 4.4499592781066895, + "step": 4275 + }, + { + "epoch": 59.391266375545854, + "grad_norm": 0.017847446724772453, + "learning_rate": 0.0006, + "loss": 4.473999500274658, + "step": 4276 + }, + { + "epoch": 59.40524017467249, + "grad_norm": 0.014944756403565407, + "learning_rate": 0.0006, + "loss": 4.5934529304504395, + "step": 4277 + }, + { + "epoch": 59.419213973799124, + "grad_norm": 0.01663832925260067, + "learning_rate": 0.0006, + "loss": 4.504262924194336, + "step": 4278 + }, + { + "epoch": 59.43318777292576, + "grad_norm": 0.015571796335279942, + "learning_rate": 0.0006, + "loss": 4.512310981750488, + "step": 4279 + }, + { + "epoch": 59.4471615720524, + "grad_norm": 0.015896400436758995, + "learning_rate": 0.0006, + "loss": 4.452714920043945, + "step": 4280 + }, + { + "epoch": 59.46113537117904, + "grad_norm": 0.016010284423828125, + "learning_rate": 0.0006, + "loss": 4.6188063621521, + "step": 4281 + }, + { + "epoch": 59.47510917030568, + "grad_norm": 0.015075593255460262, + "learning_rate": 0.0006, + "loss": 4.477978229522705, + "step": 4282 + }, + { + "epoch": 59.48908296943232, + "grad_norm": 0.014698950573801994, + "learning_rate": 0.0006, + "loss": 4.544955730438232, + "step": 4283 + }, + { + "epoch": 59.50305676855895, + "grad_norm": 0.0136245833709836, + "learning_rate": 0.0006, + "loss": 4.530460357666016, + "step": 4284 + }, + { + "epoch": 59.51703056768559, + "grad_norm": 0.013118310831487179, + "learning_rate": 0.0006, + "loss": 4.614217758178711, + "step": 4285 + }, + { + "epoch": 59.531004366812226, + "grad_norm": 0.01295961532741785, + "learning_rate": 0.0006, + "loss": 4.43156099319458, + "step": 4286 + }, + { + "epoch": 59.544978165938865, + "grad_norm": 0.012429811991751194, + "learning_rate": 0.0006, + "loss": 4.517333030700684, + "step": 4287 + }, + { + "epoch": 59.5589519650655, + "grad_norm": 0.01261830423027277, + "learning_rate": 0.0006, + "loss": 4.569780349731445, + "step": 4288 + }, + { + "epoch": 59.57292576419214, + "grad_norm": 0.012394499965012074, + "learning_rate": 0.0006, + "loss": 4.5970869064331055, + "step": 4289 + }, + { + "epoch": 59.58689956331878, + "grad_norm": 0.01290759164839983, + "learning_rate": 0.0006, + "loss": 4.437100410461426, + "step": 4290 + }, + { + "epoch": 59.60087336244541, + "grad_norm": 0.012042169459164143, + "learning_rate": 0.0006, + "loss": 4.567404747009277, + "step": 4291 + }, + { + "epoch": 59.61484716157205, + "grad_norm": 0.012338520959019661, + "learning_rate": 0.0006, + "loss": 4.476193428039551, + "step": 4292 + }, + { + "epoch": 59.62882096069869, + "grad_norm": 0.01321258582174778, + "learning_rate": 0.0006, + "loss": 4.566680431365967, + "step": 4293 + }, + { + "epoch": 59.64279475982533, + "grad_norm": 0.013478122651576996, + "learning_rate": 0.0006, + "loss": 4.487398147583008, + "step": 4294 + }, + { + "epoch": 59.65676855895197, + "grad_norm": 0.01537603884935379, + "learning_rate": 0.0006, + "loss": 4.449337005615234, + "step": 4295 + }, + { + "epoch": 59.670742358078606, + "grad_norm": 0.017140565440058708, + "learning_rate": 0.0006, + "loss": 4.5174360275268555, + "step": 4296 + }, + { + "epoch": 59.68471615720524, + "grad_norm": 0.017931461334228516, + "learning_rate": 0.0006, + "loss": 4.522104740142822, + "step": 4297 + }, + { + "epoch": 59.698689956331876, + "grad_norm": 0.018037918955087662, + "learning_rate": 0.0006, + "loss": 4.496928691864014, + "step": 4298 + }, + { + "epoch": 59.712663755458514, + "grad_norm": 0.01684504561126232, + "learning_rate": 0.0006, + "loss": 4.5332417488098145, + "step": 4299 + }, + { + "epoch": 59.72663755458515, + "grad_norm": 0.016268383711576462, + "learning_rate": 0.0006, + "loss": 4.468893051147461, + "step": 4300 + }, + { + "epoch": 59.74061135371179, + "grad_norm": 0.014859385788440704, + "learning_rate": 0.0006, + "loss": 4.551092147827148, + "step": 4301 + }, + { + "epoch": 59.75458515283843, + "grad_norm": 0.013589047826826572, + "learning_rate": 0.0006, + "loss": 4.470579147338867, + "step": 4302 + }, + { + "epoch": 59.76855895196506, + "grad_norm": 0.015704551711678505, + "learning_rate": 0.0006, + "loss": 4.330068111419678, + "step": 4303 + }, + { + "epoch": 59.7825327510917, + "grad_norm": 0.017869921401143074, + "learning_rate": 0.0006, + "loss": 4.428255081176758, + "step": 4304 + }, + { + "epoch": 59.79650655021834, + "grad_norm": 0.019828537479043007, + "learning_rate": 0.0006, + "loss": 4.496023178100586, + "step": 4305 + }, + { + "epoch": 59.81048034934498, + "grad_norm": 0.015856236219406128, + "learning_rate": 0.0006, + "loss": 4.427152156829834, + "step": 4306 + }, + { + "epoch": 59.82445414847162, + "grad_norm": 0.013349304907023907, + "learning_rate": 0.0006, + "loss": 4.417333602905273, + "step": 4307 + }, + { + "epoch": 59.838427947598255, + "grad_norm": 0.014111637137830257, + "learning_rate": 0.0006, + "loss": 4.4507575035095215, + "step": 4308 + }, + { + "epoch": 59.852401746724894, + "grad_norm": 0.015186597593128681, + "learning_rate": 0.0006, + "loss": 4.375116348266602, + "step": 4309 + }, + { + "epoch": 59.866375545851525, + "grad_norm": 0.01604730449616909, + "learning_rate": 0.0006, + "loss": 4.604518890380859, + "step": 4310 + }, + { + "epoch": 59.880349344978164, + "grad_norm": 0.015308464877307415, + "learning_rate": 0.0006, + "loss": 4.507232189178467, + "step": 4311 + }, + { + "epoch": 59.8943231441048, + "grad_norm": 0.014017206616699696, + "learning_rate": 0.0006, + "loss": 4.398487091064453, + "step": 4312 + }, + { + "epoch": 59.90829694323144, + "grad_norm": 0.012911394238471985, + "learning_rate": 0.0006, + "loss": 4.567872047424316, + "step": 4313 + }, + { + "epoch": 59.92227074235808, + "grad_norm": 0.012775142677128315, + "learning_rate": 0.0006, + "loss": 4.562105655670166, + "step": 4314 + }, + { + "epoch": 59.93624454148472, + "grad_norm": 0.014750408008694649, + "learning_rate": 0.0006, + "loss": 4.456263542175293, + "step": 4315 + }, + { + "epoch": 59.95021834061135, + "grad_norm": 0.016726011410355568, + "learning_rate": 0.0006, + "loss": 4.360722064971924, + "step": 4316 + }, + { + "epoch": 59.96419213973799, + "grad_norm": 0.01734776981174946, + "learning_rate": 0.0006, + "loss": 4.572544574737549, + "step": 4317 + }, + { + "epoch": 59.97816593886463, + "grad_norm": 0.01822415366768837, + "learning_rate": 0.0006, + "loss": 4.490501880645752, + "step": 4318 + }, + { + "epoch": 59.992139737991266, + "grad_norm": 0.014638577587902546, + "learning_rate": 0.0006, + "loss": 4.440951824188232, + "step": 4319 + }, + { + "epoch": 60.0, + "grad_norm": 0.01255202479660511, + "learning_rate": 0.0006, + "loss": 4.559123516082764, + "step": 4320 + }, + { + "epoch": 60.0, + "eval_loss": 4.741089344024658, + "eval_runtime": 56.8997, + "eval_samples_per_second": 42.918, + "eval_steps_per_second": 1.353, + "step": 4320 + }, + { + "epoch": 60.01397379912664, + "grad_norm": 0.016548393294215202, + "learning_rate": 0.0006, + "loss": 4.442024230957031, + "step": 4321 + }, + { + "epoch": 60.02794759825328, + "grad_norm": 0.019759127870202065, + "learning_rate": 0.0006, + "loss": 4.394334316253662, + "step": 4322 + }, + { + "epoch": 60.041921397379916, + "grad_norm": 0.018325170502066612, + "learning_rate": 0.0006, + "loss": 4.318270683288574, + "step": 4323 + }, + { + "epoch": 60.05589519650655, + "grad_norm": 0.015289463102817535, + "learning_rate": 0.0006, + "loss": 4.318228721618652, + "step": 4324 + }, + { + "epoch": 60.069868995633186, + "grad_norm": 0.015424872748553753, + "learning_rate": 0.0006, + "loss": 4.3605499267578125, + "step": 4325 + }, + { + "epoch": 60.083842794759825, + "grad_norm": 0.014085669070482254, + "learning_rate": 0.0006, + "loss": 4.359585762023926, + "step": 4326 + }, + { + "epoch": 60.09781659388646, + "grad_norm": 0.014004210010170937, + "learning_rate": 0.0006, + "loss": 4.447783470153809, + "step": 4327 + }, + { + "epoch": 60.1117903930131, + "grad_norm": 0.013514582999050617, + "learning_rate": 0.0006, + "loss": 4.4850239753723145, + "step": 4328 + }, + { + "epoch": 60.12576419213974, + "grad_norm": 0.013049394823610783, + "learning_rate": 0.0006, + "loss": 4.528675556182861, + "step": 4329 + }, + { + "epoch": 60.13973799126637, + "grad_norm": 0.01378058921545744, + "learning_rate": 0.0006, + "loss": 4.444823741912842, + "step": 4330 + }, + { + "epoch": 60.15371179039301, + "grad_norm": 0.013752233237028122, + "learning_rate": 0.0006, + "loss": 4.5205206871032715, + "step": 4331 + }, + { + "epoch": 60.16768558951965, + "grad_norm": 0.013431803323328495, + "learning_rate": 0.0006, + "loss": 4.547320365905762, + "step": 4332 + }, + { + "epoch": 60.18165938864629, + "grad_norm": 0.014544196426868439, + "learning_rate": 0.0006, + "loss": 4.517143249511719, + "step": 4333 + }, + { + "epoch": 60.19563318777293, + "grad_norm": 0.01612176187336445, + "learning_rate": 0.0006, + "loss": 4.517232894897461, + "step": 4334 + }, + { + "epoch": 60.209606986899566, + "grad_norm": 0.01803239807486534, + "learning_rate": 0.0006, + "loss": 4.377574920654297, + "step": 4335 + }, + { + "epoch": 60.223580786026204, + "grad_norm": 0.014871489256620407, + "learning_rate": 0.0006, + "loss": 4.538285255432129, + "step": 4336 + }, + { + "epoch": 60.237554585152836, + "grad_norm": 0.013927377760410309, + "learning_rate": 0.0006, + "loss": 4.386175632476807, + "step": 4337 + }, + { + "epoch": 60.251528384279474, + "grad_norm": 0.014757219702005386, + "learning_rate": 0.0006, + "loss": 4.528553009033203, + "step": 4338 + }, + { + "epoch": 60.26550218340611, + "grad_norm": 0.01433873176574707, + "learning_rate": 0.0006, + "loss": 4.476929187774658, + "step": 4339 + }, + { + "epoch": 60.27947598253275, + "grad_norm": 0.015768803656101227, + "learning_rate": 0.0006, + "loss": 4.50970458984375, + "step": 4340 + }, + { + "epoch": 60.29344978165939, + "grad_norm": 0.015651429072022438, + "learning_rate": 0.0006, + "loss": 4.427133083343506, + "step": 4341 + }, + { + "epoch": 60.30742358078603, + "grad_norm": 0.01673000678420067, + "learning_rate": 0.0006, + "loss": 4.447723865509033, + "step": 4342 + }, + { + "epoch": 60.32139737991266, + "grad_norm": 0.018481194972991943, + "learning_rate": 0.0006, + "loss": 4.559832572937012, + "step": 4343 + }, + { + "epoch": 60.3353711790393, + "grad_norm": 0.016731027513742447, + "learning_rate": 0.0006, + "loss": 4.470893859863281, + "step": 4344 + }, + { + "epoch": 60.34934497816594, + "grad_norm": 0.01504011545330286, + "learning_rate": 0.0006, + "loss": 4.483014106750488, + "step": 4345 + }, + { + "epoch": 60.36331877729258, + "grad_norm": 0.01602456159889698, + "learning_rate": 0.0006, + "loss": 4.548428058624268, + "step": 4346 + }, + { + "epoch": 60.377292576419215, + "grad_norm": 0.016027364879846573, + "learning_rate": 0.0006, + "loss": 4.426779747009277, + "step": 4347 + }, + { + "epoch": 60.391266375545854, + "grad_norm": 0.01873827911913395, + "learning_rate": 0.0006, + "loss": 4.353686332702637, + "step": 4348 + }, + { + "epoch": 60.40524017467249, + "grad_norm": 0.01907406374812126, + "learning_rate": 0.0006, + "loss": 4.502945423126221, + "step": 4349 + }, + { + "epoch": 60.419213973799124, + "grad_norm": 0.016083354130387306, + "learning_rate": 0.0006, + "loss": 4.431731700897217, + "step": 4350 + }, + { + "epoch": 60.43318777292576, + "grad_norm": 0.015293709933757782, + "learning_rate": 0.0006, + "loss": 4.480112552642822, + "step": 4351 + }, + { + "epoch": 60.4471615720524, + "grad_norm": 0.01688101328909397, + "learning_rate": 0.0006, + "loss": 4.50853157043457, + "step": 4352 + }, + { + "epoch": 60.46113537117904, + "grad_norm": 0.02094118855893612, + "learning_rate": 0.0006, + "loss": 4.359930038452148, + "step": 4353 + }, + { + "epoch": 60.47510917030568, + "grad_norm": 0.018188107758760452, + "learning_rate": 0.0006, + "loss": 4.549592018127441, + "step": 4354 + }, + { + "epoch": 60.48908296943232, + "grad_norm": 0.01462319865822792, + "learning_rate": 0.0006, + "loss": 4.49467658996582, + "step": 4355 + }, + { + "epoch": 60.50305676855895, + "grad_norm": 0.017508579418063164, + "learning_rate": 0.0006, + "loss": 4.398318767547607, + "step": 4356 + }, + { + "epoch": 60.51703056768559, + "grad_norm": 0.018200315535068512, + "learning_rate": 0.0006, + "loss": 4.44169282913208, + "step": 4357 + }, + { + "epoch": 60.531004366812226, + "grad_norm": 0.01717468351125717, + "learning_rate": 0.0006, + "loss": 4.481400489807129, + "step": 4358 + }, + { + "epoch": 60.544978165938865, + "grad_norm": 0.019329151138663292, + "learning_rate": 0.0006, + "loss": 4.578839302062988, + "step": 4359 + }, + { + "epoch": 60.5589519650655, + "grad_norm": 0.016109555959701538, + "learning_rate": 0.0006, + "loss": 4.461043357849121, + "step": 4360 + }, + { + "epoch": 60.57292576419214, + "grad_norm": 0.01494457945227623, + "learning_rate": 0.0006, + "loss": 4.505096912384033, + "step": 4361 + }, + { + "epoch": 60.58689956331878, + "grad_norm": 0.01729399710893631, + "learning_rate": 0.0006, + "loss": 4.366747856140137, + "step": 4362 + }, + { + "epoch": 60.60087336244541, + "grad_norm": 0.01705821603536606, + "learning_rate": 0.0006, + "loss": 4.42534065246582, + "step": 4363 + }, + { + "epoch": 60.61484716157205, + "grad_norm": 0.018391354009509087, + "learning_rate": 0.0006, + "loss": 4.350294589996338, + "step": 4364 + }, + { + "epoch": 60.62882096069869, + "grad_norm": 0.016989829018712044, + "learning_rate": 0.0006, + "loss": 4.45905876159668, + "step": 4365 + }, + { + "epoch": 60.64279475982533, + "grad_norm": 0.017192212864756584, + "learning_rate": 0.0006, + "loss": 4.526487350463867, + "step": 4366 + }, + { + "epoch": 60.65676855895197, + "grad_norm": 0.018091315403580666, + "learning_rate": 0.0006, + "loss": 4.309333324432373, + "step": 4367 + }, + { + "epoch": 60.670742358078606, + "grad_norm": 0.014051459729671478, + "learning_rate": 0.0006, + "loss": 4.338171005249023, + "step": 4368 + }, + { + "epoch": 60.68471615720524, + "grad_norm": 0.013166418299078941, + "learning_rate": 0.0006, + "loss": 4.409758567810059, + "step": 4369 + }, + { + "epoch": 60.698689956331876, + "grad_norm": 0.014260428957641125, + "learning_rate": 0.0006, + "loss": 4.524338245391846, + "step": 4370 + }, + { + "epoch": 60.712663755458514, + "grad_norm": 0.017019255086779594, + "learning_rate": 0.0006, + "loss": 4.387436866760254, + "step": 4371 + }, + { + "epoch": 60.72663755458515, + "grad_norm": 0.017070675268769264, + "learning_rate": 0.0006, + "loss": 4.437321662902832, + "step": 4372 + }, + { + "epoch": 60.74061135371179, + "grad_norm": 0.015197164379060268, + "learning_rate": 0.0006, + "loss": 4.371205806732178, + "step": 4373 + }, + { + "epoch": 60.75458515283843, + "grad_norm": 0.014791185967624187, + "learning_rate": 0.0006, + "loss": 4.44439697265625, + "step": 4374 + }, + { + "epoch": 60.76855895196506, + "grad_norm": 0.016458261758089066, + "learning_rate": 0.0006, + "loss": 4.445584297180176, + "step": 4375 + }, + { + "epoch": 60.7825327510917, + "grad_norm": 0.017706342041492462, + "learning_rate": 0.0006, + "loss": 4.348708152770996, + "step": 4376 + }, + { + "epoch": 60.79650655021834, + "grad_norm": 0.017561476677656174, + "learning_rate": 0.0006, + "loss": 4.521929740905762, + "step": 4377 + }, + { + "epoch": 60.81048034934498, + "grad_norm": 0.016557445749640465, + "learning_rate": 0.0006, + "loss": 4.545806884765625, + "step": 4378 + }, + { + "epoch": 60.82445414847162, + "grad_norm": 0.016953222453594208, + "learning_rate": 0.0006, + "loss": 4.4230875968933105, + "step": 4379 + }, + { + "epoch": 60.838427947598255, + "grad_norm": 0.017094967886805534, + "learning_rate": 0.0006, + "loss": 4.423985481262207, + "step": 4380 + }, + { + "epoch": 60.852401746724894, + "grad_norm": 0.014787515625357628, + "learning_rate": 0.0006, + "loss": 4.39473819732666, + "step": 4381 + }, + { + "epoch": 60.866375545851525, + "grad_norm": 0.016208263114094734, + "learning_rate": 0.0006, + "loss": 4.3810038566589355, + "step": 4382 + }, + { + "epoch": 60.880349344978164, + "grad_norm": 0.017844321206212044, + "learning_rate": 0.0006, + "loss": 4.35086727142334, + "step": 4383 + }, + { + "epoch": 60.8943231441048, + "grad_norm": 0.017539294436573982, + "learning_rate": 0.0006, + "loss": 4.460260391235352, + "step": 4384 + }, + { + "epoch": 60.90829694323144, + "grad_norm": 0.01612042263150215, + "learning_rate": 0.0006, + "loss": 4.370500564575195, + "step": 4385 + }, + { + "epoch": 60.92227074235808, + "grad_norm": 0.015381601639091969, + "learning_rate": 0.0006, + "loss": 4.435868263244629, + "step": 4386 + }, + { + "epoch": 60.93624454148472, + "grad_norm": 0.01603585295379162, + "learning_rate": 0.0006, + "loss": 4.434070587158203, + "step": 4387 + }, + { + "epoch": 60.95021834061135, + "grad_norm": 0.016268130391836166, + "learning_rate": 0.0006, + "loss": 4.404343128204346, + "step": 4388 + }, + { + "epoch": 60.96419213973799, + "grad_norm": 0.014980032108724117, + "learning_rate": 0.0006, + "loss": 4.399213790893555, + "step": 4389 + }, + { + "epoch": 60.97816593886463, + "grad_norm": 0.016899267211556435, + "learning_rate": 0.0006, + "loss": 4.508069038391113, + "step": 4390 + }, + { + "epoch": 60.992139737991266, + "grad_norm": 0.013756908476352692, + "learning_rate": 0.0006, + "loss": 4.541858673095703, + "step": 4391 + }, + { + "epoch": 61.0, + "grad_norm": 0.01578413136303425, + "learning_rate": 0.0006, + "loss": 4.543513298034668, + "step": 4392 + }, + { + "epoch": 61.0, + "eval_loss": 4.756770610809326, + "eval_runtime": 56.4572, + "eval_samples_per_second": 43.254, + "eval_steps_per_second": 1.364, + "step": 4392 + }, + { + "epoch": 61.01397379912664, + "grad_norm": 0.013387506827712059, + "learning_rate": 0.0006, + "loss": 4.475942611694336, + "step": 4393 + }, + { + "epoch": 61.02794759825328, + "grad_norm": 0.014278904534876347, + "learning_rate": 0.0006, + "loss": 4.442403793334961, + "step": 4394 + }, + { + "epoch": 61.041921397379916, + "grad_norm": 0.014359553344547749, + "learning_rate": 0.0006, + "loss": 4.509463310241699, + "step": 4395 + }, + { + "epoch": 61.05589519650655, + "grad_norm": 0.01718050241470337, + "learning_rate": 0.0006, + "loss": 4.354458808898926, + "step": 4396 + }, + { + "epoch": 61.069868995633186, + "grad_norm": 0.015988312661647797, + "learning_rate": 0.0006, + "loss": 4.484508514404297, + "step": 4397 + }, + { + "epoch": 61.083842794759825, + "grad_norm": 0.015466735698282719, + "learning_rate": 0.0006, + "loss": 4.326908111572266, + "step": 4398 + }, + { + "epoch": 61.09781659388646, + "grad_norm": 0.014541332609951496, + "learning_rate": 0.0006, + "loss": 4.405823230743408, + "step": 4399 + }, + { + "epoch": 61.1117903930131, + "grad_norm": 0.014803987927734852, + "learning_rate": 0.0006, + "loss": 4.375386714935303, + "step": 4400 + }, + { + "epoch": 61.12576419213974, + "grad_norm": 0.013962093740701675, + "learning_rate": 0.0006, + "loss": 4.300838470458984, + "step": 4401 + }, + { + "epoch": 61.13973799126637, + "grad_norm": 0.015213954262435436, + "learning_rate": 0.0006, + "loss": 4.388306617736816, + "step": 4402 + }, + { + "epoch": 61.15371179039301, + "grad_norm": 0.018965506926178932, + "learning_rate": 0.0006, + "loss": 4.388340950012207, + "step": 4403 + }, + { + "epoch": 61.16768558951965, + "grad_norm": 0.01958410255610943, + "learning_rate": 0.0006, + "loss": 4.35012149810791, + "step": 4404 + }, + { + "epoch": 61.18165938864629, + "grad_norm": 0.01938101276755333, + "learning_rate": 0.0006, + "loss": 4.368372917175293, + "step": 4405 + }, + { + "epoch": 61.19563318777293, + "grad_norm": 0.016556516289711, + "learning_rate": 0.0006, + "loss": 4.486198425292969, + "step": 4406 + }, + { + "epoch": 61.209606986899566, + "grad_norm": 0.017031289637088776, + "learning_rate": 0.0006, + "loss": 4.421676158905029, + "step": 4407 + }, + { + "epoch": 61.223580786026204, + "grad_norm": 0.018023479729890823, + "learning_rate": 0.0006, + "loss": 4.38751220703125, + "step": 4408 + }, + { + "epoch": 61.237554585152836, + "grad_norm": 0.0178135447204113, + "learning_rate": 0.0006, + "loss": 4.336528301239014, + "step": 4409 + }, + { + "epoch": 61.251528384279474, + "grad_norm": 0.016782190650701523, + "learning_rate": 0.0006, + "loss": 4.386693954467773, + "step": 4410 + }, + { + "epoch": 61.26550218340611, + "grad_norm": 0.015494848601520061, + "learning_rate": 0.0006, + "loss": 4.308948040008545, + "step": 4411 + }, + { + "epoch": 61.27947598253275, + "grad_norm": 0.015002378262579441, + "learning_rate": 0.0006, + "loss": 4.314206123352051, + "step": 4412 + }, + { + "epoch": 61.29344978165939, + "grad_norm": 0.015798920765519142, + "learning_rate": 0.0006, + "loss": 4.502900123596191, + "step": 4413 + }, + { + "epoch": 61.30742358078603, + "grad_norm": 0.015444336459040642, + "learning_rate": 0.0006, + "loss": 4.473160743713379, + "step": 4414 + }, + { + "epoch": 61.32139737991266, + "grad_norm": 0.017025096341967583, + "learning_rate": 0.0006, + "loss": 4.460621356964111, + "step": 4415 + }, + { + "epoch": 61.3353711790393, + "grad_norm": 0.019502537325024605, + "learning_rate": 0.0006, + "loss": 4.385551452636719, + "step": 4416 + }, + { + "epoch": 61.34934497816594, + "grad_norm": 0.018529541790485382, + "learning_rate": 0.0006, + "loss": 4.425390720367432, + "step": 4417 + }, + { + "epoch": 61.36331877729258, + "grad_norm": 0.013155706226825714, + "learning_rate": 0.0006, + "loss": 4.467020034790039, + "step": 4418 + }, + { + "epoch": 61.377292576419215, + "grad_norm": 0.01518090907484293, + "learning_rate": 0.0006, + "loss": 4.369664192199707, + "step": 4419 + }, + { + "epoch": 61.391266375545854, + "grad_norm": 0.01626511849462986, + "learning_rate": 0.0006, + "loss": 4.463231086730957, + "step": 4420 + }, + { + "epoch": 61.40524017467249, + "grad_norm": 0.01701558195054531, + "learning_rate": 0.0006, + "loss": 4.43708610534668, + "step": 4421 + }, + { + "epoch": 61.419213973799124, + "grad_norm": 0.01761862449347973, + "learning_rate": 0.0006, + "loss": 4.3314208984375, + "step": 4422 + }, + { + "epoch": 61.43318777292576, + "grad_norm": 0.017084648832678795, + "learning_rate": 0.0006, + "loss": 4.39676570892334, + "step": 4423 + }, + { + "epoch": 61.4471615720524, + "grad_norm": 0.01667078025639057, + "learning_rate": 0.0006, + "loss": 4.3789873123168945, + "step": 4424 + }, + { + "epoch": 61.46113537117904, + "grad_norm": 0.01877436414361, + "learning_rate": 0.0006, + "loss": 4.415275573730469, + "step": 4425 + }, + { + "epoch": 61.47510917030568, + "grad_norm": 0.020250486209988594, + "learning_rate": 0.0006, + "loss": 4.478635787963867, + "step": 4426 + }, + { + "epoch": 61.48908296943232, + "grad_norm": 0.020244868472218513, + "learning_rate": 0.0006, + "loss": 4.296854019165039, + "step": 4427 + }, + { + "epoch": 61.50305676855895, + "grad_norm": 0.018723875284194946, + "learning_rate": 0.0006, + "loss": 4.3760881423950195, + "step": 4428 + }, + { + "epoch": 61.51703056768559, + "grad_norm": 0.015138164162635803, + "learning_rate": 0.0006, + "loss": 4.39974308013916, + "step": 4429 + }, + { + "epoch": 61.531004366812226, + "grad_norm": 0.015111138112843037, + "learning_rate": 0.0006, + "loss": 4.34659481048584, + "step": 4430 + }, + { + "epoch": 61.544978165938865, + "grad_norm": 0.0164666585624218, + "learning_rate": 0.0006, + "loss": 4.415737152099609, + "step": 4431 + }, + { + "epoch": 61.5589519650655, + "grad_norm": 0.020891312509775162, + "learning_rate": 0.0006, + "loss": 4.300319671630859, + "step": 4432 + }, + { + "epoch": 61.57292576419214, + "grad_norm": 0.02089156024158001, + "learning_rate": 0.0006, + "loss": 4.373029708862305, + "step": 4433 + }, + { + "epoch": 61.58689956331878, + "grad_norm": 0.020114559680223465, + "learning_rate": 0.0006, + "loss": 4.434499740600586, + "step": 4434 + }, + { + "epoch": 61.60087336244541, + "grad_norm": 0.01791365258395672, + "learning_rate": 0.0006, + "loss": 4.466872215270996, + "step": 4435 + }, + { + "epoch": 61.61484716157205, + "grad_norm": 0.01879458874464035, + "learning_rate": 0.0006, + "loss": 4.5176520347595215, + "step": 4436 + }, + { + "epoch": 61.62882096069869, + "grad_norm": 0.016371937468647957, + "learning_rate": 0.0006, + "loss": 4.3507585525512695, + "step": 4437 + }, + { + "epoch": 61.64279475982533, + "grad_norm": 0.01648143120110035, + "learning_rate": 0.0006, + "loss": 4.398778438568115, + "step": 4438 + }, + { + "epoch": 61.65676855895197, + "grad_norm": 0.017240576446056366, + "learning_rate": 0.0006, + "loss": 4.513933181762695, + "step": 4439 + }, + { + "epoch": 61.670742358078606, + "grad_norm": 0.01734284684062004, + "learning_rate": 0.0006, + "loss": 4.315343856811523, + "step": 4440 + }, + { + "epoch": 61.68471615720524, + "grad_norm": 0.01639772206544876, + "learning_rate": 0.0006, + "loss": 4.4641218185424805, + "step": 4441 + }, + { + "epoch": 61.698689956331876, + "grad_norm": 0.0164708960801363, + "learning_rate": 0.0006, + "loss": 4.439968585968018, + "step": 4442 + }, + { + "epoch": 61.712663755458514, + "grad_norm": 0.01757933758199215, + "learning_rate": 0.0006, + "loss": 4.408327102661133, + "step": 4443 + }, + { + "epoch": 61.72663755458515, + "grad_norm": 0.01789589412510395, + "learning_rate": 0.0006, + "loss": 4.478257179260254, + "step": 4444 + }, + { + "epoch": 61.74061135371179, + "grad_norm": 0.01822184957563877, + "learning_rate": 0.0006, + "loss": 4.412046909332275, + "step": 4445 + }, + { + "epoch": 61.75458515283843, + "grad_norm": 0.018624618649482727, + "learning_rate": 0.0006, + "loss": 4.427122116088867, + "step": 4446 + }, + { + "epoch": 61.76855895196506, + "grad_norm": 0.018674887716770172, + "learning_rate": 0.0006, + "loss": 4.429147720336914, + "step": 4447 + }, + { + "epoch": 61.7825327510917, + "grad_norm": 0.015317119657993317, + "learning_rate": 0.0006, + "loss": 4.2808451652526855, + "step": 4448 + }, + { + "epoch": 61.79650655021834, + "grad_norm": 0.015400845557451248, + "learning_rate": 0.0006, + "loss": 4.420193672180176, + "step": 4449 + }, + { + "epoch": 61.81048034934498, + "grad_norm": 0.016743650659918785, + "learning_rate": 0.0006, + "loss": 4.498254776000977, + "step": 4450 + }, + { + "epoch": 61.82445414847162, + "grad_norm": 0.014330781064927578, + "learning_rate": 0.0006, + "loss": 4.445384979248047, + "step": 4451 + }, + { + "epoch": 61.838427947598255, + "grad_norm": 0.015046479180455208, + "learning_rate": 0.0006, + "loss": 4.368041038513184, + "step": 4452 + }, + { + "epoch": 61.852401746724894, + "grad_norm": 0.01630168780684471, + "learning_rate": 0.0006, + "loss": 4.460142135620117, + "step": 4453 + }, + { + "epoch": 61.866375545851525, + "grad_norm": 0.01637885719537735, + "learning_rate": 0.0006, + "loss": 4.416606426239014, + "step": 4454 + }, + { + "epoch": 61.880349344978164, + "grad_norm": 0.016785888001322746, + "learning_rate": 0.0006, + "loss": 4.452545642852783, + "step": 4455 + }, + { + "epoch": 61.8943231441048, + "grad_norm": 0.015044555068016052, + "learning_rate": 0.0006, + "loss": 4.427334785461426, + "step": 4456 + }, + { + "epoch": 61.90829694323144, + "grad_norm": 0.013699211180210114, + "learning_rate": 0.0006, + "loss": 4.408761024475098, + "step": 4457 + }, + { + "epoch": 61.92227074235808, + "grad_norm": 0.014021803624927998, + "learning_rate": 0.0006, + "loss": 4.543607711791992, + "step": 4458 + }, + { + "epoch": 61.93624454148472, + "grad_norm": 0.015438460744917393, + "learning_rate": 0.0006, + "loss": 4.313363075256348, + "step": 4459 + }, + { + "epoch": 61.95021834061135, + "grad_norm": 0.017515428364276886, + "learning_rate": 0.0006, + "loss": 4.310412406921387, + "step": 4460 + }, + { + "epoch": 61.96419213973799, + "grad_norm": 0.016683880239725113, + "learning_rate": 0.0006, + "loss": 4.3378448486328125, + "step": 4461 + }, + { + "epoch": 61.97816593886463, + "grad_norm": 0.019384048879146576, + "learning_rate": 0.0006, + "loss": 4.450568675994873, + "step": 4462 + }, + { + "epoch": 61.992139737991266, + "grad_norm": 0.020651493221521378, + "learning_rate": 0.0006, + "loss": 4.453197479248047, + "step": 4463 + }, + { + "epoch": 62.0, + "grad_norm": 0.020991764962673187, + "learning_rate": 0.0006, + "loss": 4.441527366638184, + "step": 4464 + }, + { + "epoch": 62.0, + "eval_loss": 4.700193881988525, + "eval_runtime": 56.6098, + "eval_samples_per_second": 43.137, + "eval_steps_per_second": 1.36, + "step": 4464 + }, + { + "epoch": 62.01397379912664, + "grad_norm": 0.01811373233795166, + "learning_rate": 0.0006, + "loss": 4.37346076965332, + "step": 4465 + }, + { + "epoch": 62.02794759825328, + "grad_norm": 0.01667127199470997, + "learning_rate": 0.0006, + "loss": 4.237151145935059, + "step": 4466 + }, + { + "epoch": 62.041921397379916, + "grad_norm": 0.014971534721553326, + "learning_rate": 0.0006, + "loss": 4.367053031921387, + "step": 4467 + }, + { + "epoch": 62.05589519650655, + "grad_norm": 0.015655893832445145, + "learning_rate": 0.0006, + "loss": 4.437426567077637, + "step": 4468 + }, + { + "epoch": 62.069868995633186, + "grad_norm": 0.014935447834432125, + "learning_rate": 0.0006, + "loss": 4.3635125160217285, + "step": 4469 + }, + { + "epoch": 62.083842794759825, + "grad_norm": 0.015363802202045918, + "learning_rate": 0.0006, + "loss": 4.40492057800293, + "step": 4470 + }, + { + "epoch": 62.09781659388646, + "grad_norm": 0.0159201230853796, + "learning_rate": 0.0006, + "loss": 4.392058849334717, + "step": 4471 + }, + { + "epoch": 62.1117903930131, + "grad_norm": 0.014598241075873375, + "learning_rate": 0.0006, + "loss": 4.369944095611572, + "step": 4472 + }, + { + "epoch": 62.12576419213974, + "grad_norm": 0.015394841320812702, + "learning_rate": 0.0006, + "loss": 4.425097942352295, + "step": 4473 + }, + { + "epoch": 62.13973799126637, + "grad_norm": 0.015913628041744232, + "learning_rate": 0.0006, + "loss": 4.420596122741699, + "step": 4474 + }, + { + "epoch": 62.15371179039301, + "grad_norm": 0.01575258933007717, + "learning_rate": 0.0006, + "loss": 4.32290506362915, + "step": 4475 + }, + { + "epoch": 62.16768558951965, + "grad_norm": 0.0155344782397151, + "learning_rate": 0.0006, + "loss": 4.273068428039551, + "step": 4476 + }, + { + "epoch": 62.18165938864629, + "grad_norm": 0.014947040006518364, + "learning_rate": 0.0006, + "loss": 4.384868621826172, + "step": 4477 + }, + { + "epoch": 62.19563318777293, + "grad_norm": 0.01557591650635004, + "learning_rate": 0.0006, + "loss": 4.458410263061523, + "step": 4478 + }, + { + "epoch": 62.209606986899566, + "grad_norm": 0.015182198956608772, + "learning_rate": 0.0006, + "loss": 4.300738334655762, + "step": 4479 + }, + { + "epoch": 62.223580786026204, + "grad_norm": 0.01634717360138893, + "learning_rate": 0.0006, + "loss": 4.436875820159912, + "step": 4480 + }, + { + "epoch": 62.237554585152836, + "grad_norm": 0.018813718110322952, + "learning_rate": 0.0006, + "loss": 4.454665184020996, + "step": 4481 + }, + { + "epoch": 62.251528384279474, + "grad_norm": 0.020965630188584328, + "learning_rate": 0.0006, + "loss": 4.357341766357422, + "step": 4482 + }, + { + "epoch": 62.26550218340611, + "grad_norm": 0.020738353952765465, + "learning_rate": 0.0006, + "loss": 4.456644058227539, + "step": 4483 + }, + { + "epoch": 62.27947598253275, + "grad_norm": 0.018813546746969223, + "learning_rate": 0.0006, + "loss": 4.381941795349121, + "step": 4484 + }, + { + "epoch": 62.29344978165939, + "grad_norm": 0.015431761741638184, + "learning_rate": 0.0006, + "loss": 4.41126823425293, + "step": 4485 + }, + { + "epoch": 62.30742358078603, + "grad_norm": 0.016335977241396904, + "learning_rate": 0.0006, + "loss": 4.419488906860352, + "step": 4486 + }, + { + "epoch": 62.32139737991266, + "grad_norm": 0.01832406409084797, + "learning_rate": 0.0006, + "loss": 4.579026222229004, + "step": 4487 + }, + { + "epoch": 62.3353711790393, + "grad_norm": 0.01790742017328739, + "learning_rate": 0.0006, + "loss": 4.289964199066162, + "step": 4488 + }, + { + "epoch": 62.34934497816594, + "grad_norm": 0.017266202718019485, + "learning_rate": 0.0006, + "loss": 4.415959358215332, + "step": 4489 + }, + { + "epoch": 62.36331877729258, + "grad_norm": 0.015976132825016975, + "learning_rate": 0.0006, + "loss": 4.358587265014648, + "step": 4490 + }, + { + "epoch": 62.377292576419215, + "grad_norm": 0.016873881220817566, + "learning_rate": 0.0006, + "loss": 4.354607105255127, + "step": 4491 + }, + { + "epoch": 62.391266375545854, + "grad_norm": 0.01476745493710041, + "learning_rate": 0.0006, + "loss": 4.384904384613037, + "step": 4492 + }, + { + "epoch": 62.40524017467249, + "grad_norm": 0.013810782693326473, + "learning_rate": 0.0006, + "loss": 4.459230899810791, + "step": 4493 + }, + { + "epoch": 62.419213973799124, + "grad_norm": 0.015150421299040318, + "learning_rate": 0.0006, + "loss": 4.369471549987793, + "step": 4494 + }, + { + "epoch": 62.43318777292576, + "grad_norm": 0.014539610594511032, + "learning_rate": 0.0006, + "loss": 4.3564982414245605, + "step": 4495 + }, + { + "epoch": 62.4471615720524, + "grad_norm": 0.015554017387330532, + "learning_rate": 0.0006, + "loss": 4.414379119873047, + "step": 4496 + }, + { + "epoch": 62.46113537117904, + "grad_norm": 0.016608357429504395, + "learning_rate": 0.0006, + "loss": 4.311585426330566, + "step": 4497 + }, + { + "epoch": 62.47510917030568, + "grad_norm": 0.01885881833732128, + "learning_rate": 0.0006, + "loss": 4.306629180908203, + "step": 4498 + }, + { + "epoch": 62.48908296943232, + "grad_norm": 0.022128663957118988, + "learning_rate": 0.0006, + "loss": 4.345483779907227, + "step": 4499 + }, + { + "epoch": 62.50305676855895, + "grad_norm": 0.023401018232107162, + "learning_rate": 0.0006, + "loss": 4.398204326629639, + "step": 4500 + }, + { + "epoch": 62.51703056768559, + "grad_norm": 0.022080009803175926, + "learning_rate": 0.0006, + "loss": 4.375027179718018, + "step": 4501 + }, + { + "epoch": 62.531004366812226, + "grad_norm": 0.020004864782094955, + "learning_rate": 0.0006, + "loss": 4.33494758605957, + "step": 4502 + }, + { + "epoch": 62.544978165938865, + "grad_norm": 0.01786700263619423, + "learning_rate": 0.0006, + "loss": 4.4428300857543945, + "step": 4503 + }, + { + "epoch": 62.5589519650655, + "grad_norm": 0.01885388419032097, + "learning_rate": 0.0006, + "loss": 4.4138946533203125, + "step": 4504 + }, + { + "epoch": 62.57292576419214, + "grad_norm": 0.019890939816832542, + "learning_rate": 0.0006, + "loss": 4.278632164001465, + "step": 4505 + }, + { + "epoch": 62.58689956331878, + "grad_norm": 0.02109031192958355, + "learning_rate": 0.0006, + "loss": 4.3756327629089355, + "step": 4506 + }, + { + "epoch": 62.60087336244541, + "grad_norm": 0.020813899114727974, + "learning_rate": 0.0006, + "loss": 4.410213470458984, + "step": 4507 + }, + { + "epoch": 62.61484716157205, + "grad_norm": 0.021655689924955368, + "learning_rate": 0.0006, + "loss": 4.30633020401001, + "step": 4508 + }, + { + "epoch": 62.62882096069869, + "grad_norm": 0.018155649304389954, + "learning_rate": 0.0006, + "loss": 4.407284736633301, + "step": 4509 + }, + { + "epoch": 62.64279475982533, + "grad_norm": 0.017765069380402565, + "learning_rate": 0.0006, + "loss": 4.321715354919434, + "step": 4510 + }, + { + "epoch": 62.65676855895197, + "grad_norm": 0.017724554985761642, + "learning_rate": 0.0006, + "loss": 4.493371963500977, + "step": 4511 + }, + { + "epoch": 62.670742358078606, + "grad_norm": 0.01492092851549387, + "learning_rate": 0.0006, + "loss": 4.443025588989258, + "step": 4512 + }, + { + "epoch": 62.68471615720524, + "grad_norm": 0.01615324057638645, + "learning_rate": 0.0006, + "loss": 4.3289594650268555, + "step": 4513 + }, + { + "epoch": 62.698689956331876, + "grad_norm": 0.016569366678595543, + "learning_rate": 0.0006, + "loss": 4.369720458984375, + "step": 4514 + }, + { + "epoch": 62.712663755458514, + "grad_norm": 0.016509901732206345, + "learning_rate": 0.0006, + "loss": 4.292316436767578, + "step": 4515 + }, + { + "epoch": 62.72663755458515, + "grad_norm": 0.014270003885030746, + "learning_rate": 0.0006, + "loss": 4.349854469299316, + "step": 4516 + }, + { + "epoch": 62.74061135371179, + "grad_norm": 0.014235340990126133, + "learning_rate": 0.0006, + "loss": 4.473727226257324, + "step": 4517 + }, + { + "epoch": 62.75458515283843, + "grad_norm": 0.015163225121796131, + "learning_rate": 0.0006, + "loss": 4.452878475189209, + "step": 4518 + }, + { + "epoch": 62.76855895196506, + "grad_norm": 0.015748055651783943, + "learning_rate": 0.0006, + "loss": 4.521087169647217, + "step": 4519 + }, + { + "epoch": 62.7825327510917, + "grad_norm": 0.01661108434200287, + "learning_rate": 0.0006, + "loss": 4.443885326385498, + "step": 4520 + }, + { + "epoch": 62.79650655021834, + "grad_norm": 0.01610301062464714, + "learning_rate": 0.0006, + "loss": 4.312990188598633, + "step": 4521 + }, + { + "epoch": 62.81048034934498, + "grad_norm": 0.017007583752274513, + "learning_rate": 0.0006, + "loss": 4.408937454223633, + "step": 4522 + }, + { + "epoch": 62.82445414847162, + "grad_norm": 0.01749461516737938, + "learning_rate": 0.0006, + "loss": 4.346066951751709, + "step": 4523 + }, + { + "epoch": 62.838427947598255, + "grad_norm": 0.018243493512272835, + "learning_rate": 0.0006, + "loss": 4.357658386230469, + "step": 4524 + }, + { + "epoch": 62.852401746724894, + "grad_norm": 0.017499985173344612, + "learning_rate": 0.0006, + "loss": 4.374061584472656, + "step": 4525 + }, + { + "epoch": 62.866375545851525, + "grad_norm": 0.014536325819790363, + "learning_rate": 0.0006, + "loss": 4.189321994781494, + "step": 4526 + }, + { + "epoch": 62.880349344978164, + "grad_norm": 0.014728891663253307, + "learning_rate": 0.0006, + "loss": 4.425910949707031, + "step": 4527 + }, + { + "epoch": 62.8943231441048, + "grad_norm": 0.015013772062957287, + "learning_rate": 0.0006, + "loss": 4.337319850921631, + "step": 4528 + }, + { + "epoch": 62.90829694323144, + "grad_norm": 0.01626027375459671, + "learning_rate": 0.0006, + "loss": 4.330339431762695, + "step": 4529 + }, + { + "epoch": 62.92227074235808, + "grad_norm": 0.014773444272577763, + "learning_rate": 0.0006, + "loss": 4.4188737869262695, + "step": 4530 + }, + { + "epoch": 62.93624454148472, + "grad_norm": 0.01422932744026184, + "learning_rate": 0.0006, + "loss": 4.355986595153809, + "step": 4531 + }, + { + "epoch": 62.95021834061135, + "grad_norm": 0.012170110829174519, + "learning_rate": 0.0006, + "loss": 4.405722618103027, + "step": 4532 + }, + { + "epoch": 62.96419213973799, + "grad_norm": 0.013233253732323647, + "learning_rate": 0.0006, + "loss": 4.377286911010742, + "step": 4533 + }, + { + "epoch": 62.97816593886463, + "grad_norm": 0.014497299678623676, + "learning_rate": 0.0006, + "loss": 4.508796691894531, + "step": 4534 + }, + { + "epoch": 62.992139737991266, + "grad_norm": 0.014718741178512573, + "learning_rate": 0.0006, + "loss": 4.370919704437256, + "step": 4535 + }, + { + "epoch": 63.0, + "grad_norm": 0.015833502635359764, + "learning_rate": 0.0006, + "loss": 4.258753776550293, + "step": 4536 + }, + { + "epoch": 63.0, + "eval_loss": 4.714163303375244, + "eval_runtime": 57.0269, + "eval_samples_per_second": 42.822, + "eval_steps_per_second": 1.35, + "step": 4536 + }, + { + "epoch": 63.01397379912664, + "grad_norm": 0.014277510344982147, + "learning_rate": 0.0006, + "loss": 4.3112640380859375, + "step": 4537 + }, + { + "epoch": 63.02794759825328, + "grad_norm": 0.015377935022115707, + "learning_rate": 0.0006, + "loss": 4.342782974243164, + "step": 4538 + }, + { + "epoch": 63.041921397379916, + "grad_norm": 0.015406244434416294, + "learning_rate": 0.0006, + "loss": 4.474056720733643, + "step": 4539 + }, + { + "epoch": 63.05589519650655, + "grad_norm": 0.018121499568223953, + "learning_rate": 0.0006, + "loss": 4.397963523864746, + "step": 4540 + }, + { + "epoch": 63.069868995633186, + "grad_norm": 0.022634636610746384, + "learning_rate": 0.0006, + "loss": 4.306631088256836, + "step": 4541 + }, + { + "epoch": 63.083842794759825, + "grad_norm": 0.026448970660567284, + "learning_rate": 0.0006, + "loss": 4.283175468444824, + "step": 4542 + }, + { + "epoch": 63.09781659388646, + "grad_norm": 0.024104537442326546, + "learning_rate": 0.0006, + "loss": 4.315826416015625, + "step": 4543 + }, + { + "epoch": 63.1117903930131, + "grad_norm": 0.019040480256080627, + "learning_rate": 0.0006, + "loss": 4.44057035446167, + "step": 4544 + }, + { + "epoch": 63.12576419213974, + "grad_norm": 0.02006000280380249, + "learning_rate": 0.0006, + "loss": 4.405603408813477, + "step": 4545 + }, + { + "epoch": 63.13973799126637, + "grad_norm": 0.019980071112513542, + "learning_rate": 0.0006, + "loss": 4.393726348876953, + "step": 4546 + }, + { + "epoch": 63.15371179039301, + "grad_norm": 0.01641865260899067, + "learning_rate": 0.0006, + "loss": 4.272546768188477, + "step": 4547 + }, + { + "epoch": 63.16768558951965, + "grad_norm": 0.01890353485941887, + "learning_rate": 0.0006, + "loss": 4.382566928863525, + "step": 4548 + }, + { + "epoch": 63.18165938864629, + "grad_norm": 0.018370937556028366, + "learning_rate": 0.0006, + "loss": 4.423953533172607, + "step": 4549 + }, + { + "epoch": 63.19563318777293, + "grad_norm": 0.014985294081270695, + "learning_rate": 0.0006, + "loss": 4.308627128601074, + "step": 4550 + }, + { + "epoch": 63.209606986899566, + "grad_norm": 0.017513327300548553, + "learning_rate": 0.0006, + "loss": 4.248819351196289, + "step": 4551 + }, + { + "epoch": 63.223580786026204, + "grad_norm": 0.019464161247015, + "learning_rate": 0.0006, + "loss": 4.359729766845703, + "step": 4552 + }, + { + "epoch": 63.237554585152836, + "grad_norm": 0.018678035587072372, + "learning_rate": 0.0006, + "loss": 4.3299665451049805, + "step": 4553 + }, + { + "epoch": 63.251528384279474, + "grad_norm": 0.017909523099660873, + "learning_rate": 0.0006, + "loss": 4.246147155761719, + "step": 4554 + }, + { + "epoch": 63.26550218340611, + "grad_norm": 0.01747293211519718, + "learning_rate": 0.0006, + "loss": 4.41165018081665, + "step": 4555 + }, + { + "epoch": 63.27947598253275, + "grad_norm": 0.018670853227376938, + "learning_rate": 0.0006, + "loss": 4.409883499145508, + "step": 4556 + }, + { + "epoch": 63.29344978165939, + "grad_norm": 0.0189529862254858, + "learning_rate": 0.0006, + "loss": 4.418654918670654, + "step": 4557 + }, + { + "epoch": 63.30742358078603, + "grad_norm": 0.018394414335489273, + "learning_rate": 0.0006, + "loss": 4.38355827331543, + "step": 4558 + }, + { + "epoch": 63.32139737991266, + "grad_norm": 0.018317820504307747, + "learning_rate": 0.0006, + "loss": 4.456843376159668, + "step": 4559 + }, + { + "epoch": 63.3353711790393, + "grad_norm": 0.019225597381591797, + "learning_rate": 0.0006, + "loss": 4.232264518737793, + "step": 4560 + }, + { + "epoch": 63.34934497816594, + "grad_norm": 0.016663681715726852, + "learning_rate": 0.0006, + "loss": 4.393517017364502, + "step": 4561 + }, + { + "epoch": 63.36331877729258, + "grad_norm": 0.014975365251302719, + "learning_rate": 0.0006, + "loss": 4.375183582305908, + "step": 4562 + }, + { + "epoch": 63.377292576419215, + "grad_norm": 0.01771732047200203, + "learning_rate": 0.0006, + "loss": 4.3073410987854, + "step": 4563 + }, + { + "epoch": 63.391266375545854, + "grad_norm": 0.01663549244403839, + "learning_rate": 0.0006, + "loss": 4.311003684997559, + "step": 4564 + }, + { + "epoch": 63.40524017467249, + "grad_norm": 0.016301177442073822, + "learning_rate": 0.0006, + "loss": 4.326807022094727, + "step": 4565 + }, + { + "epoch": 63.419213973799124, + "grad_norm": 0.016781188547611237, + "learning_rate": 0.0006, + "loss": 4.306562423706055, + "step": 4566 + }, + { + "epoch": 63.43318777292576, + "grad_norm": 0.01767745055258274, + "learning_rate": 0.0006, + "loss": 4.367082595825195, + "step": 4567 + }, + { + "epoch": 63.4471615720524, + "grad_norm": 0.016728565096855164, + "learning_rate": 0.0006, + "loss": 4.327400207519531, + "step": 4568 + }, + { + "epoch": 63.46113537117904, + "grad_norm": 0.016289403662085533, + "learning_rate": 0.0006, + "loss": 4.392712116241455, + "step": 4569 + }, + { + "epoch": 63.47510917030568, + "grad_norm": 0.016233716160058975, + "learning_rate": 0.0006, + "loss": 4.354372024536133, + "step": 4570 + }, + { + "epoch": 63.48908296943232, + "grad_norm": 0.016547974199056625, + "learning_rate": 0.0006, + "loss": 4.309050559997559, + "step": 4571 + }, + { + "epoch": 63.50305676855895, + "grad_norm": 0.018466509878635406, + "learning_rate": 0.0006, + "loss": 4.318345069885254, + "step": 4572 + }, + { + "epoch": 63.51703056768559, + "grad_norm": 0.017685720697045326, + "learning_rate": 0.0006, + "loss": 4.399726867675781, + "step": 4573 + }, + { + "epoch": 63.531004366812226, + "grad_norm": 0.015801234170794487, + "learning_rate": 0.0006, + "loss": 4.328495502471924, + "step": 4574 + }, + { + "epoch": 63.544978165938865, + "grad_norm": 0.0188873540610075, + "learning_rate": 0.0006, + "loss": 4.253244400024414, + "step": 4575 + }, + { + "epoch": 63.5589519650655, + "grad_norm": 0.01868271268904209, + "learning_rate": 0.0006, + "loss": 4.463746070861816, + "step": 4576 + }, + { + "epoch": 63.57292576419214, + "grad_norm": 0.01844949834048748, + "learning_rate": 0.0006, + "loss": 4.400300025939941, + "step": 4577 + }, + { + "epoch": 63.58689956331878, + "grad_norm": 0.015256480313837528, + "learning_rate": 0.0006, + "loss": 4.445844650268555, + "step": 4578 + }, + { + "epoch": 63.60087336244541, + "grad_norm": 0.015305760316550732, + "learning_rate": 0.0006, + "loss": 4.317327499389648, + "step": 4579 + }, + { + "epoch": 63.61484716157205, + "grad_norm": 0.015200769528746605, + "learning_rate": 0.0006, + "loss": 4.25112247467041, + "step": 4580 + }, + { + "epoch": 63.62882096069869, + "grad_norm": 0.015919538214802742, + "learning_rate": 0.0006, + "loss": 4.277937889099121, + "step": 4581 + }, + { + "epoch": 63.64279475982533, + "grad_norm": 0.01572665013372898, + "learning_rate": 0.0006, + "loss": 4.302237510681152, + "step": 4582 + }, + { + "epoch": 63.65676855895197, + "grad_norm": 0.016428731381893158, + "learning_rate": 0.0006, + "loss": 4.36762809753418, + "step": 4583 + }, + { + "epoch": 63.670742358078606, + "grad_norm": 0.01706082746386528, + "learning_rate": 0.0006, + "loss": 4.236526966094971, + "step": 4584 + }, + { + "epoch": 63.68471615720524, + "grad_norm": 0.017850136384367943, + "learning_rate": 0.0006, + "loss": 4.264129638671875, + "step": 4585 + }, + { + "epoch": 63.698689956331876, + "grad_norm": 0.016079438850283623, + "learning_rate": 0.0006, + "loss": 4.449799060821533, + "step": 4586 + }, + { + "epoch": 63.712663755458514, + "grad_norm": 0.014996383339166641, + "learning_rate": 0.0006, + "loss": 4.404574394226074, + "step": 4587 + }, + { + "epoch": 63.72663755458515, + "grad_norm": 0.015910431742668152, + "learning_rate": 0.0006, + "loss": 4.348101615905762, + "step": 4588 + }, + { + "epoch": 63.74061135371179, + "grad_norm": 0.01630021072924137, + "learning_rate": 0.0006, + "loss": 4.3193817138671875, + "step": 4589 + }, + { + "epoch": 63.75458515283843, + "grad_norm": 0.014571773819625378, + "learning_rate": 0.0006, + "loss": 4.374402046203613, + "step": 4590 + }, + { + "epoch": 63.76855895196506, + "grad_norm": 0.01662345975637436, + "learning_rate": 0.0006, + "loss": 4.329075813293457, + "step": 4591 + }, + { + "epoch": 63.7825327510917, + "grad_norm": 0.016604198142886162, + "learning_rate": 0.0006, + "loss": 4.436126232147217, + "step": 4592 + }, + { + "epoch": 63.79650655021834, + "grad_norm": 0.016578884795308113, + "learning_rate": 0.0006, + "loss": 4.361468315124512, + "step": 4593 + }, + { + "epoch": 63.81048034934498, + "grad_norm": 0.016339825466275215, + "learning_rate": 0.0006, + "loss": 4.449892044067383, + "step": 4594 + }, + { + "epoch": 63.82445414847162, + "grad_norm": 0.016805455088615417, + "learning_rate": 0.0006, + "loss": 4.415866851806641, + "step": 4595 + }, + { + "epoch": 63.838427947598255, + "grad_norm": 0.017366213724017143, + "learning_rate": 0.0006, + "loss": 4.361730575561523, + "step": 4596 + }, + { + "epoch": 63.852401746724894, + "grad_norm": 0.015599003992974758, + "learning_rate": 0.0006, + "loss": 4.363637924194336, + "step": 4597 + }, + { + "epoch": 63.866375545851525, + "grad_norm": 0.015674632042646408, + "learning_rate": 0.0006, + "loss": 4.332045078277588, + "step": 4598 + }, + { + "epoch": 63.880349344978164, + "grad_norm": 0.015350443311035633, + "learning_rate": 0.0006, + "loss": 4.320063591003418, + "step": 4599 + }, + { + "epoch": 63.8943231441048, + "grad_norm": 0.015388146974146366, + "learning_rate": 0.0006, + "loss": 4.398199081420898, + "step": 4600 + }, + { + "epoch": 63.90829694323144, + "grad_norm": 0.01450140681117773, + "learning_rate": 0.0006, + "loss": 4.244076728820801, + "step": 4601 + }, + { + "epoch": 63.92227074235808, + "grad_norm": 0.01637491025030613, + "learning_rate": 0.0006, + "loss": 4.213549613952637, + "step": 4602 + }, + { + "epoch": 63.93624454148472, + "grad_norm": 0.0159814041107893, + "learning_rate": 0.0006, + "loss": 4.421775817871094, + "step": 4603 + }, + { + "epoch": 63.95021834061135, + "grad_norm": 0.017386827617883682, + "learning_rate": 0.0006, + "loss": 4.340023994445801, + "step": 4604 + }, + { + "epoch": 63.96419213973799, + "grad_norm": 0.019679656252264977, + "learning_rate": 0.0006, + "loss": 4.352534294128418, + "step": 4605 + }, + { + "epoch": 63.97816593886463, + "grad_norm": 0.022391516715288162, + "learning_rate": 0.0006, + "loss": 4.3538312911987305, + "step": 4606 + }, + { + "epoch": 63.992139737991266, + "grad_norm": 0.023281006142497063, + "learning_rate": 0.0006, + "loss": 4.432089328765869, + "step": 4607 + }, + { + "epoch": 64.0, + "grad_norm": 0.019916288554668427, + "learning_rate": 0.0006, + "loss": 4.466937065124512, + "step": 4608 + }, + { + "epoch": 64.0, + "eval_loss": 4.659008502960205, + "eval_runtime": 56.0019, + "eval_samples_per_second": 43.606, + "eval_steps_per_second": 1.375, + "step": 4608 + }, + { + "epoch": 64.01397379912663, + "grad_norm": 0.017351187765598297, + "learning_rate": 0.0006, + "loss": 4.294129371643066, + "step": 4609 + }, + { + "epoch": 64.02794759825328, + "grad_norm": 0.02405865117907524, + "learning_rate": 0.0006, + "loss": 4.266717910766602, + "step": 4610 + }, + { + "epoch": 64.04192139737991, + "grad_norm": 0.022158386185765266, + "learning_rate": 0.0006, + "loss": 4.395932197570801, + "step": 4611 + }, + { + "epoch": 64.05589519650655, + "grad_norm": 0.019238505512475967, + "learning_rate": 0.0006, + "loss": 4.266704082489014, + "step": 4612 + }, + { + "epoch": 64.06986899563319, + "grad_norm": 0.021670425310730934, + "learning_rate": 0.0006, + "loss": 4.3160552978515625, + "step": 4613 + }, + { + "epoch": 64.08384279475983, + "grad_norm": 0.02076641097664833, + "learning_rate": 0.0006, + "loss": 4.415626525878906, + "step": 4614 + }, + { + "epoch": 64.09781659388646, + "grad_norm": 0.018223078921437263, + "learning_rate": 0.0006, + "loss": 4.292387008666992, + "step": 4615 + }, + { + "epoch": 64.1117903930131, + "grad_norm": 0.017401937395334244, + "learning_rate": 0.0006, + "loss": 4.424524307250977, + "step": 4616 + }, + { + "epoch": 64.12576419213974, + "grad_norm": 0.01663743332028389, + "learning_rate": 0.0006, + "loss": 4.350032806396484, + "step": 4617 + }, + { + "epoch": 64.13973799126637, + "grad_norm": 0.018716629594564438, + "learning_rate": 0.0006, + "loss": 4.309157371520996, + "step": 4618 + }, + { + "epoch": 64.15371179039302, + "grad_norm": 0.019569825381040573, + "learning_rate": 0.0006, + "loss": 4.358292579650879, + "step": 4619 + }, + { + "epoch": 64.16768558951965, + "grad_norm": 0.018994422629475594, + "learning_rate": 0.0006, + "loss": 4.3338823318481445, + "step": 4620 + }, + { + "epoch": 64.18165938864628, + "grad_norm": 0.017690688371658325, + "learning_rate": 0.0006, + "loss": 4.414468765258789, + "step": 4621 + }, + { + "epoch": 64.19563318777293, + "grad_norm": 0.0170609001070261, + "learning_rate": 0.0006, + "loss": 4.4337568283081055, + "step": 4622 + }, + { + "epoch": 64.20960698689956, + "grad_norm": 0.016984131187200546, + "learning_rate": 0.0006, + "loss": 4.304839134216309, + "step": 4623 + }, + { + "epoch": 64.2235807860262, + "grad_norm": 0.01619867794215679, + "learning_rate": 0.0006, + "loss": 4.389623165130615, + "step": 4624 + }, + { + "epoch": 64.23755458515284, + "grad_norm": 0.015428110957145691, + "learning_rate": 0.0006, + "loss": 4.277515888214111, + "step": 4625 + }, + { + "epoch": 64.25152838427948, + "grad_norm": 0.015909532085061073, + "learning_rate": 0.0006, + "loss": 4.266029357910156, + "step": 4626 + }, + { + "epoch": 64.26550218340611, + "grad_norm": 0.014296560548245907, + "learning_rate": 0.0006, + "loss": 4.286067485809326, + "step": 4627 + }, + { + "epoch": 64.27947598253274, + "grad_norm": 0.014065971598029137, + "learning_rate": 0.0006, + "loss": 4.353528022766113, + "step": 4628 + }, + { + "epoch": 64.29344978165939, + "grad_norm": 0.014386293478310108, + "learning_rate": 0.0006, + "loss": 4.318715572357178, + "step": 4629 + }, + { + "epoch": 64.30742358078602, + "grad_norm": 0.014001123607158661, + "learning_rate": 0.0006, + "loss": 4.362497329711914, + "step": 4630 + }, + { + "epoch": 64.32139737991267, + "grad_norm": 0.013408979400992393, + "learning_rate": 0.0006, + "loss": 4.417397499084473, + "step": 4631 + }, + { + "epoch": 64.3353711790393, + "grad_norm": 0.014960093423724174, + "learning_rate": 0.0006, + "loss": 4.359710693359375, + "step": 4632 + }, + { + "epoch": 64.34934497816595, + "grad_norm": 0.014658777043223381, + "learning_rate": 0.0006, + "loss": 4.318822383880615, + "step": 4633 + }, + { + "epoch": 64.36331877729258, + "grad_norm": 0.013766744174063206, + "learning_rate": 0.0006, + "loss": 4.265679836273193, + "step": 4634 + }, + { + "epoch": 64.37729257641921, + "grad_norm": 0.013821255415678024, + "learning_rate": 0.0006, + "loss": 4.3452911376953125, + "step": 4635 + }, + { + "epoch": 64.39126637554585, + "grad_norm": 0.013541066087782383, + "learning_rate": 0.0006, + "loss": 4.429254531860352, + "step": 4636 + }, + { + "epoch": 64.40524017467249, + "grad_norm": 0.013985645025968552, + "learning_rate": 0.0006, + "loss": 4.304452896118164, + "step": 4637 + }, + { + "epoch": 64.41921397379913, + "grad_norm": 0.015079808421432972, + "learning_rate": 0.0006, + "loss": 4.296828746795654, + "step": 4638 + }, + { + "epoch": 64.43318777292576, + "grad_norm": 0.014896944165229797, + "learning_rate": 0.0006, + "loss": 4.299156665802002, + "step": 4639 + }, + { + "epoch": 64.44716157205241, + "grad_norm": 0.013770041987299919, + "learning_rate": 0.0006, + "loss": 4.351439952850342, + "step": 4640 + }, + { + "epoch": 64.46113537117904, + "grad_norm": 0.01531178504228592, + "learning_rate": 0.0006, + "loss": 4.373715877532959, + "step": 4641 + }, + { + "epoch": 64.47510917030567, + "grad_norm": 0.014024450443685055, + "learning_rate": 0.0006, + "loss": 4.368053436279297, + "step": 4642 + }, + { + "epoch": 64.48908296943232, + "grad_norm": 0.013864024542272091, + "learning_rate": 0.0006, + "loss": 4.378564834594727, + "step": 4643 + }, + { + "epoch": 64.50305676855895, + "grad_norm": 0.013925445266067982, + "learning_rate": 0.0006, + "loss": 4.261663436889648, + "step": 4644 + }, + { + "epoch": 64.5170305676856, + "grad_norm": 0.015011338517069817, + "learning_rate": 0.0006, + "loss": 4.337285995483398, + "step": 4645 + }, + { + "epoch": 64.53100436681223, + "grad_norm": 0.016941986978054047, + "learning_rate": 0.0006, + "loss": 4.336877822875977, + "step": 4646 + }, + { + "epoch": 64.54497816593886, + "grad_norm": 0.01719909906387329, + "learning_rate": 0.0006, + "loss": 4.149813175201416, + "step": 4647 + }, + { + "epoch": 64.5589519650655, + "grad_norm": 0.01608402095735073, + "learning_rate": 0.0006, + "loss": 4.191280364990234, + "step": 4648 + }, + { + "epoch": 64.57292576419214, + "grad_norm": 0.015284820459783077, + "learning_rate": 0.0006, + "loss": 4.35953426361084, + "step": 4649 + }, + { + "epoch": 64.58689956331878, + "grad_norm": 0.014855682849884033, + "learning_rate": 0.0006, + "loss": 4.308401107788086, + "step": 4650 + }, + { + "epoch": 64.60087336244541, + "grad_norm": 0.017553111538290977, + "learning_rate": 0.0006, + "loss": 4.279720306396484, + "step": 4651 + }, + { + "epoch": 64.61484716157206, + "grad_norm": 0.019506998360157013, + "learning_rate": 0.0006, + "loss": 4.319524765014648, + "step": 4652 + }, + { + "epoch": 64.62882096069869, + "grad_norm": 0.020293425768613815, + "learning_rate": 0.0006, + "loss": 4.340523719787598, + "step": 4653 + }, + { + "epoch": 64.64279475982532, + "grad_norm": 0.01664556749165058, + "learning_rate": 0.0006, + "loss": 4.296785831451416, + "step": 4654 + }, + { + "epoch": 64.65676855895197, + "grad_norm": 0.012592585757374763, + "learning_rate": 0.0006, + "loss": 4.464595317840576, + "step": 4655 + }, + { + "epoch": 64.6707423580786, + "grad_norm": 0.01549589168280363, + "learning_rate": 0.0006, + "loss": 4.275053024291992, + "step": 4656 + }, + { + "epoch": 64.68471615720524, + "grad_norm": 0.01705172471702099, + "learning_rate": 0.0006, + "loss": 4.388431549072266, + "step": 4657 + }, + { + "epoch": 64.69868995633188, + "grad_norm": 0.015438582748174667, + "learning_rate": 0.0006, + "loss": 4.312070846557617, + "step": 4658 + }, + { + "epoch": 64.71266375545852, + "grad_norm": 0.014448083005845547, + "learning_rate": 0.0006, + "loss": 4.237599849700928, + "step": 4659 + }, + { + "epoch": 64.72663755458515, + "grad_norm": 0.015267071314156055, + "learning_rate": 0.0006, + "loss": 4.29893684387207, + "step": 4660 + }, + { + "epoch": 64.74061135371178, + "grad_norm": 0.016615847125649452, + "learning_rate": 0.0006, + "loss": 4.432273864746094, + "step": 4661 + }, + { + "epoch": 64.75458515283843, + "grad_norm": 0.016584362834692, + "learning_rate": 0.0006, + "loss": 4.202266216278076, + "step": 4662 + }, + { + "epoch": 64.76855895196506, + "grad_norm": 0.018313994631171227, + "learning_rate": 0.0006, + "loss": 4.367966651916504, + "step": 4663 + }, + { + "epoch": 64.78253275109171, + "grad_norm": 0.017946481704711914, + "learning_rate": 0.0006, + "loss": 4.205172538757324, + "step": 4664 + }, + { + "epoch": 64.79650655021834, + "grad_norm": 0.017744354903697968, + "learning_rate": 0.0006, + "loss": 4.295661926269531, + "step": 4665 + }, + { + "epoch": 64.81048034934499, + "grad_norm": 0.016103824600577354, + "learning_rate": 0.0006, + "loss": 4.276543617248535, + "step": 4666 + }, + { + "epoch": 64.82445414847162, + "grad_norm": 0.014590012840926647, + "learning_rate": 0.0006, + "loss": 4.414247035980225, + "step": 4667 + }, + { + "epoch": 64.83842794759825, + "grad_norm": 0.01782386749982834, + "learning_rate": 0.0006, + "loss": 4.203220367431641, + "step": 4668 + }, + { + "epoch": 64.8524017467249, + "grad_norm": 0.022122733294963837, + "learning_rate": 0.0006, + "loss": 4.38078498840332, + "step": 4669 + }, + { + "epoch": 64.86637554585153, + "grad_norm": 0.0250293780118227, + "learning_rate": 0.0006, + "loss": 4.3310136795043945, + "step": 4670 + }, + { + "epoch": 64.88034934497817, + "grad_norm": 0.022438062354922295, + "learning_rate": 0.0006, + "loss": 4.428631782531738, + "step": 4671 + }, + { + "epoch": 64.8943231441048, + "grad_norm": 0.018518192693591118, + "learning_rate": 0.0006, + "loss": 4.428250312805176, + "step": 4672 + }, + { + "epoch": 64.90829694323143, + "grad_norm": 0.018105225637555122, + "learning_rate": 0.0006, + "loss": 4.336209297180176, + "step": 4673 + }, + { + "epoch": 64.92227074235808, + "grad_norm": 0.017564399167895317, + "learning_rate": 0.0006, + "loss": 4.416011810302734, + "step": 4674 + }, + { + "epoch": 64.93624454148471, + "grad_norm": 0.018747031688690186, + "learning_rate": 0.0006, + "loss": 4.419938087463379, + "step": 4675 + }, + { + "epoch": 64.95021834061136, + "grad_norm": 0.018686484545469284, + "learning_rate": 0.0006, + "loss": 4.347185134887695, + "step": 4676 + }, + { + "epoch": 64.96419213973799, + "grad_norm": 0.019329898059368134, + "learning_rate": 0.0006, + "loss": 4.343749046325684, + "step": 4677 + }, + { + "epoch": 64.97816593886463, + "grad_norm": 0.018517782911658287, + "learning_rate": 0.0006, + "loss": 4.264604568481445, + "step": 4678 + }, + { + "epoch": 64.99213973799127, + "grad_norm": 0.016579125076532364, + "learning_rate": 0.0006, + "loss": 4.303423881530762, + "step": 4679 + }, + { + "epoch": 65.0, + "grad_norm": 0.01617567613720894, + "learning_rate": 0.0006, + "loss": 4.326556205749512, + "step": 4680 + }, + { + "epoch": 65.0, + "eval_loss": 4.655026912689209, + "eval_runtime": 57.1299, + "eval_samples_per_second": 42.745, + "eval_steps_per_second": 1.348, + "step": 4680 + }, + { + "epoch": 65.01397379912663, + "grad_norm": 0.01607249490916729, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 4681 + }, + { + "epoch": 65.02794759825328, + "grad_norm": 0.016779497265815735, + "learning_rate": 0.0006, + "loss": 4.412464618682861, + "step": 4682 + }, + { + "epoch": 65.04192139737991, + "grad_norm": 0.018161555752158165, + "learning_rate": 0.0006, + "loss": 4.199329376220703, + "step": 4683 + }, + { + "epoch": 65.05589519650655, + "grad_norm": 0.017508309334516525, + "learning_rate": 0.0006, + "loss": 4.217567443847656, + "step": 4684 + }, + { + "epoch": 65.06986899563319, + "grad_norm": 0.015716979280114174, + "learning_rate": 0.0006, + "loss": 4.322023391723633, + "step": 4685 + }, + { + "epoch": 65.08384279475983, + "grad_norm": 0.015113635919988155, + "learning_rate": 0.0006, + "loss": 4.333515167236328, + "step": 4686 + }, + { + "epoch": 65.09781659388646, + "grad_norm": 0.01437560748308897, + "learning_rate": 0.0006, + "loss": 4.295609474182129, + "step": 4687 + }, + { + "epoch": 65.1117903930131, + "grad_norm": 0.013074862770736217, + "learning_rate": 0.0006, + "loss": 4.314120769500732, + "step": 4688 + }, + { + "epoch": 65.12576419213974, + "grad_norm": 0.012908284552395344, + "learning_rate": 0.0006, + "loss": 4.216890811920166, + "step": 4689 + }, + { + "epoch": 65.13973799126637, + "grad_norm": 0.013001754879951477, + "learning_rate": 0.0006, + "loss": 4.287717342376709, + "step": 4690 + }, + { + "epoch": 65.15371179039302, + "grad_norm": 0.013694263994693756, + "learning_rate": 0.0006, + "loss": 4.333837985992432, + "step": 4691 + }, + { + "epoch": 65.16768558951965, + "grad_norm": 0.014461833983659744, + "learning_rate": 0.0006, + "loss": 4.323545455932617, + "step": 4692 + }, + { + "epoch": 65.18165938864628, + "grad_norm": 0.013670675456523895, + "learning_rate": 0.0006, + "loss": 4.3313493728637695, + "step": 4693 + }, + { + "epoch": 65.19563318777293, + "grad_norm": 0.013094350695610046, + "learning_rate": 0.0006, + "loss": 4.23858118057251, + "step": 4694 + }, + { + "epoch": 65.20960698689956, + "grad_norm": 0.013996944762766361, + "learning_rate": 0.0006, + "loss": 4.428705215454102, + "step": 4695 + }, + { + "epoch": 65.2235807860262, + "grad_norm": 0.013295048847794533, + "learning_rate": 0.0006, + "loss": 4.238950252532959, + "step": 4696 + }, + { + "epoch": 65.23755458515284, + "grad_norm": 0.013380616903305054, + "learning_rate": 0.0006, + "loss": 4.273185729980469, + "step": 4697 + }, + { + "epoch": 65.25152838427948, + "grad_norm": 0.015062890015542507, + "learning_rate": 0.0006, + "loss": 4.29472017288208, + "step": 4698 + }, + { + "epoch": 65.26550218340611, + "grad_norm": 0.017317263409495354, + "learning_rate": 0.0006, + "loss": 4.337954998016357, + "step": 4699 + }, + { + "epoch": 65.27947598253274, + "grad_norm": 0.017698992043733597, + "learning_rate": 0.0006, + "loss": 4.336437225341797, + "step": 4700 + }, + { + "epoch": 65.29344978165939, + "grad_norm": 0.01628575474023819, + "learning_rate": 0.0006, + "loss": 4.302372455596924, + "step": 4701 + }, + { + "epoch": 65.30742358078602, + "grad_norm": 0.014531854540109634, + "learning_rate": 0.0006, + "loss": 4.256450653076172, + "step": 4702 + }, + { + "epoch": 65.32139737991267, + "grad_norm": 0.014549105428159237, + "learning_rate": 0.0006, + "loss": 4.302240371704102, + "step": 4703 + }, + { + "epoch": 65.3353711790393, + "grad_norm": 0.014881663955748081, + "learning_rate": 0.0006, + "loss": 4.3316121101379395, + "step": 4704 + }, + { + "epoch": 65.34934497816595, + "grad_norm": 0.014513437636196613, + "learning_rate": 0.0006, + "loss": 4.441212177276611, + "step": 4705 + }, + { + "epoch": 65.36331877729258, + "grad_norm": 0.014122087508440018, + "learning_rate": 0.0006, + "loss": 4.343296527862549, + "step": 4706 + }, + { + "epoch": 65.37729257641921, + "grad_norm": 0.014817348681390285, + "learning_rate": 0.0006, + "loss": 4.195674896240234, + "step": 4707 + }, + { + "epoch": 65.39126637554585, + "grad_norm": 0.01787056028842926, + "learning_rate": 0.0006, + "loss": 4.3507914543151855, + "step": 4708 + }, + { + "epoch": 65.40524017467249, + "grad_norm": 0.015295136719942093, + "learning_rate": 0.0006, + "loss": 4.234372138977051, + "step": 4709 + }, + { + "epoch": 65.41921397379913, + "grad_norm": 0.014324234798550606, + "learning_rate": 0.0006, + "loss": 4.2907867431640625, + "step": 4710 + }, + { + "epoch": 65.43318777292576, + "grad_norm": 0.016090670600533485, + "learning_rate": 0.0006, + "loss": 4.177244186401367, + "step": 4711 + }, + { + "epoch": 65.44716157205241, + "grad_norm": 0.0174604132771492, + "learning_rate": 0.0006, + "loss": 4.335577964782715, + "step": 4712 + }, + { + "epoch": 65.46113537117904, + "grad_norm": 0.016731899231672287, + "learning_rate": 0.0006, + "loss": 4.30197811126709, + "step": 4713 + }, + { + "epoch": 65.47510917030567, + "grad_norm": 0.017423735931515694, + "learning_rate": 0.0006, + "loss": 4.3528852462768555, + "step": 4714 + }, + { + "epoch": 65.48908296943232, + "grad_norm": 0.01780976541340351, + "learning_rate": 0.0006, + "loss": 4.185036659240723, + "step": 4715 + }, + { + "epoch": 65.50305676855895, + "grad_norm": 0.019977200776338577, + "learning_rate": 0.0006, + "loss": 4.328766345977783, + "step": 4716 + }, + { + "epoch": 65.5170305676856, + "grad_norm": 0.020284287631511688, + "learning_rate": 0.0006, + "loss": 4.285494804382324, + "step": 4717 + }, + { + "epoch": 65.53100436681223, + "grad_norm": 0.019168002530932426, + "learning_rate": 0.0006, + "loss": 4.328560829162598, + "step": 4718 + }, + { + "epoch": 65.54497816593886, + "grad_norm": 0.019217776134610176, + "learning_rate": 0.0006, + "loss": 4.189998626708984, + "step": 4719 + }, + { + "epoch": 65.5589519650655, + "grad_norm": 0.018960488960146904, + "learning_rate": 0.0006, + "loss": 4.4995012283325195, + "step": 4720 + }, + { + "epoch": 65.57292576419214, + "grad_norm": 0.01941288262605667, + "learning_rate": 0.0006, + "loss": 4.26570987701416, + "step": 4721 + }, + { + "epoch": 65.58689956331878, + "grad_norm": 0.020263448357582092, + "learning_rate": 0.0006, + "loss": 4.411645889282227, + "step": 4722 + }, + { + "epoch": 65.60087336244541, + "grad_norm": 0.02144799567759037, + "learning_rate": 0.0006, + "loss": 4.163675308227539, + "step": 4723 + }, + { + "epoch": 65.61484716157206, + "grad_norm": 0.019790446385741234, + "learning_rate": 0.0006, + "loss": 4.2361063957214355, + "step": 4724 + }, + { + "epoch": 65.62882096069869, + "grad_norm": 0.01829805225133896, + "learning_rate": 0.0006, + "loss": 4.449810028076172, + "step": 4725 + }, + { + "epoch": 65.64279475982532, + "grad_norm": 0.01571105420589447, + "learning_rate": 0.0006, + "loss": 4.281406879425049, + "step": 4726 + }, + { + "epoch": 65.65676855895197, + "grad_norm": 0.016477147117257118, + "learning_rate": 0.0006, + "loss": 4.293408393859863, + "step": 4727 + }, + { + "epoch": 65.6707423580786, + "grad_norm": 0.016766032204031944, + "learning_rate": 0.0006, + "loss": 4.378252983093262, + "step": 4728 + }, + { + "epoch": 65.68471615720524, + "grad_norm": 0.0159933939576149, + "learning_rate": 0.0006, + "loss": 4.3065667152404785, + "step": 4729 + }, + { + "epoch": 65.69868995633188, + "grad_norm": 0.01658209227025509, + "learning_rate": 0.0006, + "loss": 4.174026966094971, + "step": 4730 + }, + { + "epoch": 65.71266375545852, + "grad_norm": 0.015081380493938923, + "learning_rate": 0.0006, + "loss": 4.360222339630127, + "step": 4731 + }, + { + "epoch": 65.72663755458515, + "grad_norm": 0.015110890381038189, + "learning_rate": 0.0006, + "loss": 4.218810558319092, + "step": 4732 + }, + { + "epoch": 65.74061135371178, + "grad_norm": 0.015300953760743141, + "learning_rate": 0.0006, + "loss": 4.331979274749756, + "step": 4733 + }, + { + "epoch": 65.75458515283843, + "grad_norm": 0.01715504191815853, + "learning_rate": 0.0006, + "loss": 4.349112033843994, + "step": 4734 + }, + { + "epoch": 65.76855895196506, + "grad_norm": 0.015870366245508194, + "learning_rate": 0.0006, + "loss": 4.329850196838379, + "step": 4735 + }, + { + "epoch": 65.78253275109171, + "grad_norm": 0.01541937980800867, + "learning_rate": 0.0006, + "loss": 4.3400163650512695, + "step": 4736 + }, + { + "epoch": 65.79650655021834, + "grad_norm": 0.01738942787051201, + "learning_rate": 0.0006, + "loss": 4.220798492431641, + "step": 4737 + }, + { + "epoch": 65.81048034934499, + "grad_norm": 0.01685990020632744, + "learning_rate": 0.0006, + "loss": 4.177238464355469, + "step": 4738 + }, + { + "epoch": 65.82445414847162, + "grad_norm": 0.016183465719223022, + "learning_rate": 0.0006, + "loss": 4.327216625213623, + "step": 4739 + }, + { + "epoch": 65.83842794759825, + "grad_norm": 0.01677694357931614, + "learning_rate": 0.0006, + "loss": 4.27766227722168, + "step": 4740 + }, + { + "epoch": 65.8524017467249, + "grad_norm": 0.01645066775381565, + "learning_rate": 0.0006, + "loss": 4.288347244262695, + "step": 4741 + }, + { + "epoch": 65.86637554585153, + "grad_norm": 0.018034635111689568, + "learning_rate": 0.0006, + "loss": 4.292458534240723, + "step": 4742 + }, + { + "epoch": 65.88034934497817, + "grad_norm": 0.017973346635699272, + "learning_rate": 0.0006, + "loss": 4.3117265701293945, + "step": 4743 + }, + { + "epoch": 65.8943231441048, + "grad_norm": 0.016605669632554054, + "learning_rate": 0.0006, + "loss": 4.280615329742432, + "step": 4744 + }, + { + "epoch": 65.90829694323143, + "grad_norm": 0.016298074275255203, + "learning_rate": 0.0006, + "loss": 4.301933288574219, + "step": 4745 + }, + { + "epoch": 65.92227074235808, + "grad_norm": 0.015383216552436352, + "learning_rate": 0.0006, + "loss": 4.352383136749268, + "step": 4746 + }, + { + "epoch": 65.93624454148471, + "grad_norm": 0.016222182661294937, + "learning_rate": 0.0006, + "loss": 4.305725574493408, + "step": 4747 + }, + { + "epoch": 65.95021834061136, + "grad_norm": 0.01691329851746559, + "learning_rate": 0.0006, + "loss": 4.29502010345459, + "step": 4748 + }, + { + "epoch": 65.96419213973799, + "grad_norm": 0.017637677490711212, + "learning_rate": 0.0006, + "loss": 4.3133039474487305, + "step": 4749 + }, + { + "epoch": 65.97816593886463, + "grad_norm": 0.021362047642469406, + "learning_rate": 0.0006, + "loss": 4.458834648132324, + "step": 4750 + }, + { + "epoch": 65.99213973799127, + "grad_norm": 0.023553011938929558, + "learning_rate": 0.0006, + "loss": 4.298727989196777, + "step": 4751 + }, + { + "epoch": 66.0, + "grad_norm": 0.02154950052499771, + "learning_rate": 0.0006, + "loss": 4.403392791748047, + "step": 4752 + }, + { + "epoch": 66.0, + "eval_loss": 4.724340438842773, + "eval_runtime": 56.9555, + "eval_samples_per_second": 42.876, + "eval_steps_per_second": 1.352, + "step": 4752 + }, + { + "epoch": 66.01397379912663, + "grad_norm": 0.01916784606873989, + "learning_rate": 0.0006, + "loss": 4.22931432723999, + "step": 4753 + }, + { + "epoch": 66.02794759825328, + "grad_norm": 0.017492054030299187, + "learning_rate": 0.0006, + "loss": 4.271522521972656, + "step": 4754 + }, + { + "epoch": 66.04192139737991, + "grad_norm": 0.018408095464110374, + "learning_rate": 0.0006, + "loss": 4.337705135345459, + "step": 4755 + }, + { + "epoch": 66.05589519650655, + "grad_norm": 0.019220108166337013, + "learning_rate": 0.0006, + "loss": 4.280340194702148, + "step": 4756 + }, + { + "epoch": 66.06986899563319, + "grad_norm": 0.01773735322058201, + "learning_rate": 0.0006, + "loss": 4.390201568603516, + "step": 4757 + }, + { + "epoch": 66.08384279475983, + "grad_norm": 0.01679125428199768, + "learning_rate": 0.0006, + "loss": 4.299282073974609, + "step": 4758 + }, + { + "epoch": 66.09781659388646, + "grad_norm": 0.01479059923440218, + "learning_rate": 0.0006, + "loss": 4.472630500793457, + "step": 4759 + }, + { + "epoch": 66.1117903930131, + "grad_norm": 0.015130104497075081, + "learning_rate": 0.0006, + "loss": 4.301194190979004, + "step": 4760 + }, + { + "epoch": 66.12576419213974, + "grad_norm": 0.0158979594707489, + "learning_rate": 0.0006, + "loss": 4.375751972198486, + "step": 4761 + }, + { + "epoch": 66.13973799126637, + "grad_norm": 0.014432408846914768, + "learning_rate": 0.0006, + "loss": 4.279788017272949, + "step": 4762 + }, + { + "epoch": 66.15371179039302, + "grad_norm": 0.013026784174144268, + "learning_rate": 0.0006, + "loss": 4.255035400390625, + "step": 4763 + }, + { + "epoch": 66.16768558951965, + "grad_norm": 0.014422653242945671, + "learning_rate": 0.0006, + "loss": 4.261832237243652, + "step": 4764 + }, + { + "epoch": 66.18165938864628, + "grad_norm": 0.014613019302487373, + "learning_rate": 0.0006, + "loss": 4.382184028625488, + "step": 4765 + }, + { + "epoch": 66.19563318777293, + "grad_norm": 0.014989894814789295, + "learning_rate": 0.0006, + "loss": 4.40528678894043, + "step": 4766 + }, + { + "epoch": 66.20960698689956, + "grad_norm": 0.018190350383520126, + "learning_rate": 0.0006, + "loss": 4.302414894104004, + "step": 4767 + }, + { + "epoch": 66.2235807860262, + "grad_norm": 0.020887890830636024, + "learning_rate": 0.0006, + "loss": 4.289920806884766, + "step": 4768 + }, + { + "epoch": 66.23755458515284, + "grad_norm": 0.02376876398921013, + "learning_rate": 0.0006, + "loss": 4.297855377197266, + "step": 4769 + }, + { + "epoch": 66.25152838427948, + "grad_norm": 0.02152269519865513, + "learning_rate": 0.0006, + "loss": 4.1972832679748535, + "step": 4770 + }, + { + "epoch": 66.26550218340611, + "grad_norm": 0.015572802163660526, + "learning_rate": 0.0006, + "loss": 4.307487487792969, + "step": 4771 + }, + { + "epoch": 66.27947598253274, + "grad_norm": 0.015228749252855778, + "learning_rate": 0.0006, + "loss": 4.103250503540039, + "step": 4772 + }, + { + "epoch": 66.29344978165939, + "grad_norm": 0.016925504431128502, + "learning_rate": 0.0006, + "loss": 4.247461318969727, + "step": 4773 + }, + { + "epoch": 66.30742358078602, + "grad_norm": 0.01613711006939411, + "learning_rate": 0.0006, + "loss": 4.4699273109436035, + "step": 4774 + }, + { + "epoch": 66.32139737991267, + "grad_norm": 0.018332140520215034, + "learning_rate": 0.0006, + "loss": 4.22969388961792, + "step": 4775 + }, + { + "epoch": 66.3353711790393, + "grad_norm": 0.019575074315071106, + "learning_rate": 0.0006, + "loss": 4.29397439956665, + "step": 4776 + }, + { + "epoch": 66.34934497816595, + "grad_norm": 0.018177419900894165, + "learning_rate": 0.0006, + "loss": 4.221899509429932, + "step": 4777 + }, + { + "epoch": 66.36331877729258, + "grad_norm": 0.014292889274656773, + "learning_rate": 0.0006, + "loss": 4.325457572937012, + "step": 4778 + }, + { + "epoch": 66.37729257641921, + "grad_norm": 0.015098796226084232, + "learning_rate": 0.0006, + "loss": 4.364837169647217, + "step": 4779 + }, + { + "epoch": 66.39126637554585, + "grad_norm": 0.013992635533213615, + "learning_rate": 0.0006, + "loss": 4.299082279205322, + "step": 4780 + }, + { + "epoch": 66.40524017467249, + "grad_norm": 0.01504511572420597, + "learning_rate": 0.0006, + "loss": 4.230877876281738, + "step": 4781 + }, + { + "epoch": 66.41921397379913, + "grad_norm": 0.014875639230012894, + "learning_rate": 0.0006, + "loss": 4.368405818939209, + "step": 4782 + }, + { + "epoch": 66.43318777292576, + "grad_norm": 0.014306887984275818, + "learning_rate": 0.0006, + "loss": 4.389184474945068, + "step": 4783 + }, + { + "epoch": 66.44716157205241, + "grad_norm": 0.012959563173353672, + "learning_rate": 0.0006, + "loss": 4.37934684753418, + "step": 4784 + }, + { + "epoch": 66.46113537117904, + "grad_norm": 0.013206703588366508, + "learning_rate": 0.0006, + "loss": 4.442799091339111, + "step": 4785 + }, + { + "epoch": 66.47510917030567, + "grad_norm": 0.014848909340798855, + "learning_rate": 0.0006, + "loss": 4.203481674194336, + "step": 4786 + }, + { + "epoch": 66.48908296943232, + "grad_norm": 0.015637682750821114, + "learning_rate": 0.0006, + "loss": 4.27552604675293, + "step": 4787 + }, + { + "epoch": 66.50305676855895, + "grad_norm": 0.016017816960811615, + "learning_rate": 0.0006, + "loss": 4.296402931213379, + "step": 4788 + }, + { + "epoch": 66.5170305676856, + "grad_norm": 0.01577814109623432, + "learning_rate": 0.0006, + "loss": 4.301733016967773, + "step": 4789 + }, + { + "epoch": 66.53100436681223, + "grad_norm": 0.015293175354599953, + "learning_rate": 0.0006, + "loss": 4.369877815246582, + "step": 4790 + }, + { + "epoch": 66.54497816593886, + "grad_norm": 0.014262278564274311, + "learning_rate": 0.0006, + "loss": 4.3314738273620605, + "step": 4791 + }, + { + "epoch": 66.5589519650655, + "grad_norm": 0.015123482793569565, + "learning_rate": 0.0006, + "loss": 4.313591957092285, + "step": 4792 + }, + { + "epoch": 66.57292576419214, + "grad_norm": 0.015458766371011734, + "learning_rate": 0.0006, + "loss": 4.3045854568481445, + "step": 4793 + }, + { + "epoch": 66.58689956331878, + "grad_norm": 0.017217233777046204, + "learning_rate": 0.0006, + "loss": 4.269944190979004, + "step": 4794 + }, + { + "epoch": 66.60087336244541, + "grad_norm": 0.01962282694876194, + "learning_rate": 0.0006, + "loss": 4.271735191345215, + "step": 4795 + }, + { + "epoch": 66.61484716157206, + "grad_norm": 0.01917066052556038, + "learning_rate": 0.0006, + "loss": 4.288168430328369, + "step": 4796 + }, + { + "epoch": 66.62882096069869, + "grad_norm": 0.018211524933576584, + "learning_rate": 0.0006, + "loss": 4.280660629272461, + "step": 4797 + }, + { + "epoch": 66.64279475982532, + "grad_norm": 0.01936524175107479, + "learning_rate": 0.0006, + "loss": 4.3710551261901855, + "step": 4798 + }, + { + "epoch": 66.65676855895197, + "grad_norm": 0.01882351003587246, + "learning_rate": 0.0006, + "loss": 4.342750549316406, + "step": 4799 + }, + { + "epoch": 66.6707423580786, + "grad_norm": 0.017149122431874275, + "learning_rate": 0.0006, + "loss": 4.32134485244751, + "step": 4800 + }, + { + "epoch": 66.68471615720524, + "grad_norm": 0.01699327677488327, + "learning_rate": 0.0006, + "loss": 4.248893737792969, + "step": 4801 + }, + { + "epoch": 66.69868995633188, + "grad_norm": 0.016873255372047424, + "learning_rate": 0.0006, + "loss": 4.277350425720215, + "step": 4802 + }, + { + "epoch": 66.71266375545852, + "grad_norm": 0.016592837870121002, + "learning_rate": 0.0006, + "loss": 4.221665859222412, + "step": 4803 + }, + { + "epoch": 66.72663755458515, + "grad_norm": 0.015887131914496422, + "learning_rate": 0.0006, + "loss": 4.328040599822998, + "step": 4804 + }, + { + "epoch": 66.74061135371178, + "grad_norm": 0.014057286083698273, + "learning_rate": 0.0006, + "loss": 4.3294548988342285, + "step": 4805 + }, + { + "epoch": 66.75458515283843, + "grad_norm": 0.013135905377566814, + "learning_rate": 0.0006, + "loss": 4.254383563995361, + "step": 4806 + }, + { + "epoch": 66.76855895196506, + "grad_norm": 0.013632557354867458, + "learning_rate": 0.0006, + "loss": 4.342815399169922, + "step": 4807 + }, + { + "epoch": 66.78253275109171, + "grad_norm": 0.015512688085436821, + "learning_rate": 0.0006, + "loss": 4.297835350036621, + "step": 4808 + }, + { + "epoch": 66.79650655021834, + "grad_norm": 0.01621832884848118, + "learning_rate": 0.0006, + "loss": 4.314773082733154, + "step": 4809 + }, + { + "epoch": 66.81048034934499, + "grad_norm": 0.016196032986044884, + "learning_rate": 0.0006, + "loss": 4.29209566116333, + "step": 4810 + }, + { + "epoch": 66.82445414847162, + "grad_norm": 0.015463353134691715, + "learning_rate": 0.0006, + "loss": 4.262795448303223, + "step": 4811 + }, + { + "epoch": 66.83842794759825, + "grad_norm": 0.016143787652254105, + "learning_rate": 0.0006, + "loss": 4.274284362792969, + "step": 4812 + }, + { + "epoch": 66.8524017467249, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.337160587310791, + "step": 4813 + }, + { + "epoch": 66.86637554585153, + "grad_norm": 0.01582838036119938, + "learning_rate": 0.0006, + "loss": 4.192723274230957, + "step": 4814 + }, + { + "epoch": 66.88034934497817, + "grad_norm": 0.015081540681421757, + "learning_rate": 0.0006, + "loss": 4.370708465576172, + "step": 4815 + }, + { + "epoch": 66.8943231441048, + "grad_norm": 0.015929723158478737, + "learning_rate": 0.0006, + "loss": 4.340244293212891, + "step": 4816 + }, + { + "epoch": 66.90829694323143, + "grad_norm": 0.01784711331129074, + "learning_rate": 0.0006, + "loss": 4.260406970977783, + "step": 4817 + }, + { + "epoch": 66.92227074235808, + "grad_norm": 0.017434703186154366, + "learning_rate": 0.0006, + "loss": 4.286900043487549, + "step": 4818 + }, + { + "epoch": 66.93624454148471, + "grad_norm": 0.016262540593743324, + "learning_rate": 0.0006, + "loss": 4.292142391204834, + "step": 4819 + }, + { + "epoch": 66.95021834061136, + "grad_norm": 0.016122829169034958, + "learning_rate": 0.0006, + "loss": 4.291485786437988, + "step": 4820 + }, + { + "epoch": 66.96419213973799, + "grad_norm": 0.014361178502440453, + "learning_rate": 0.0006, + "loss": 4.275425910949707, + "step": 4821 + }, + { + "epoch": 66.97816593886463, + "grad_norm": 0.01373987551778555, + "learning_rate": 0.0006, + "loss": 4.223120212554932, + "step": 4822 + }, + { + "epoch": 66.99213973799127, + "grad_norm": 0.015687376260757446, + "learning_rate": 0.0006, + "loss": 4.355381011962891, + "step": 4823 + }, + { + "epoch": 67.0, + "grad_norm": 0.01998465321958065, + "learning_rate": 0.0006, + "loss": 4.264053821563721, + "step": 4824 + }, + { + "epoch": 67.0, + "eval_loss": 4.662398338317871, + "eval_runtime": 56.7928, + "eval_samples_per_second": 42.998, + "eval_steps_per_second": 1.356, + "step": 4824 + }, + { + "epoch": 67.01397379912663, + "grad_norm": 0.021736836060881615, + "learning_rate": 0.0006, + "loss": 4.243354797363281, + "step": 4825 + }, + { + "epoch": 67.02794759825328, + "grad_norm": 0.021476037800312042, + "learning_rate": 0.0006, + "loss": 4.218207836151123, + "step": 4826 + }, + { + "epoch": 67.04192139737991, + "grad_norm": 0.017188388854265213, + "learning_rate": 0.0006, + "loss": 4.18048095703125, + "step": 4827 + }, + { + "epoch": 67.05589519650655, + "grad_norm": 0.017293719574809074, + "learning_rate": 0.0006, + "loss": 4.239145278930664, + "step": 4828 + }, + { + "epoch": 67.06986899563319, + "grad_norm": 0.017383819445967674, + "learning_rate": 0.0006, + "loss": 4.114765167236328, + "step": 4829 + }, + { + "epoch": 67.08384279475983, + "grad_norm": 0.01651529036462307, + "learning_rate": 0.0006, + "loss": 4.351023197174072, + "step": 4830 + }, + { + "epoch": 67.09781659388646, + "grad_norm": 0.015893027186393738, + "learning_rate": 0.0006, + "loss": 4.261432647705078, + "step": 4831 + }, + { + "epoch": 67.1117903930131, + "grad_norm": 0.013381517492234707, + "learning_rate": 0.0006, + "loss": 4.172064781188965, + "step": 4832 + }, + { + "epoch": 67.12576419213974, + "grad_norm": 0.014642654918134212, + "learning_rate": 0.0006, + "loss": 4.356171131134033, + "step": 4833 + }, + { + "epoch": 67.13973799126637, + "grad_norm": 0.01425282284617424, + "learning_rate": 0.0006, + "loss": 4.303495407104492, + "step": 4834 + }, + { + "epoch": 67.15371179039302, + "grad_norm": 0.014574599452316761, + "learning_rate": 0.0006, + "loss": 4.2850847244262695, + "step": 4835 + }, + { + "epoch": 67.16768558951965, + "grad_norm": 0.012676282785832882, + "learning_rate": 0.0006, + "loss": 4.297845840454102, + "step": 4836 + }, + { + "epoch": 67.18165938864628, + "grad_norm": 0.013948136940598488, + "learning_rate": 0.0006, + "loss": 4.324530601501465, + "step": 4837 + }, + { + "epoch": 67.19563318777293, + "grad_norm": 0.014195773750543594, + "learning_rate": 0.0006, + "loss": 4.31236457824707, + "step": 4838 + }, + { + "epoch": 67.20960698689956, + "grad_norm": 0.013631382025778294, + "learning_rate": 0.0006, + "loss": 4.086362838745117, + "step": 4839 + }, + { + "epoch": 67.2235807860262, + "grad_norm": 0.012798587791621685, + "learning_rate": 0.0006, + "loss": 4.328471660614014, + "step": 4840 + }, + { + "epoch": 67.23755458515284, + "grad_norm": 0.014443321153521538, + "learning_rate": 0.0006, + "loss": 4.278019905090332, + "step": 4841 + }, + { + "epoch": 67.25152838427948, + "grad_norm": 0.015355420298874378, + "learning_rate": 0.0006, + "loss": 4.333611011505127, + "step": 4842 + }, + { + "epoch": 67.26550218340611, + "grad_norm": 0.016487296670675278, + "learning_rate": 0.0006, + "loss": 4.351336479187012, + "step": 4843 + }, + { + "epoch": 67.27947598253274, + "grad_norm": 0.017422541975975037, + "learning_rate": 0.0006, + "loss": 4.386096477508545, + "step": 4844 + }, + { + "epoch": 67.29344978165939, + "grad_norm": 0.015790918841958046, + "learning_rate": 0.0006, + "loss": 4.301377773284912, + "step": 4845 + }, + { + "epoch": 67.30742358078602, + "grad_norm": 0.016427364200353622, + "learning_rate": 0.0006, + "loss": 4.2838945388793945, + "step": 4846 + }, + { + "epoch": 67.32139737991267, + "grad_norm": 0.017050761729478836, + "learning_rate": 0.0006, + "loss": 4.282417297363281, + "step": 4847 + }, + { + "epoch": 67.3353711790393, + "grad_norm": 0.015006253495812416, + "learning_rate": 0.0006, + "loss": 4.352236270904541, + "step": 4848 + }, + { + "epoch": 67.34934497816595, + "grad_norm": 0.012996627949178219, + "learning_rate": 0.0006, + "loss": 4.309242248535156, + "step": 4849 + }, + { + "epoch": 67.36331877729258, + "grad_norm": 0.01367615070194006, + "learning_rate": 0.0006, + "loss": 4.211801528930664, + "step": 4850 + }, + { + "epoch": 67.37729257641921, + "grad_norm": 0.014775963500142097, + "learning_rate": 0.0006, + "loss": 4.360859394073486, + "step": 4851 + }, + { + "epoch": 67.39126637554585, + "grad_norm": 0.015854304656386375, + "learning_rate": 0.0006, + "loss": 4.211830139160156, + "step": 4852 + }, + { + "epoch": 67.40524017467249, + "grad_norm": 0.016888894140720367, + "learning_rate": 0.0006, + "loss": 4.216549873352051, + "step": 4853 + }, + { + "epoch": 67.41921397379913, + "grad_norm": 0.0170894768089056, + "learning_rate": 0.0006, + "loss": 4.263486385345459, + "step": 4854 + }, + { + "epoch": 67.43318777292576, + "grad_norm": 0.015829868614673615, + "learning_rate": 0.0006, + "loss": 4.360196113586426, + "step": 4855 + }, + { + "epoch": 67.44716157205241, + "grad_norm": 0.015220202505588531, + "learning_rate": 0.0006, + "loss": 4.243794918060303, + "step": 4856 + }, + { + "epoch": 67.46113537117904, + "grad_norm": 0.01413388829678297, + "learning_rate": 0.0006, + "loss": 4.286624908447266, + "step": 4857 + }, + { + "epoch": 67.47510917030567, + "grad_norm": 0.01426626369357109, + "learning_rate": 0.0006, + "loss": 4.257316589355469, + "step": 4858 + }, + { + "epoch": 67.48908296943232, + "grad_norm": 0.015957845374941826, + "learning_rate": 0.0006, + "loss": 4.300443649291992, + "step": 4859 + }, + { + "epoch": 67.50305676855895, + "grad_norm": 0.015499051660299301, + "learning_rate": 0.0006, + "loss": 4.278162002563477, + "step": 4860 + }, + { + "epoch": 67.5170305676856, + "grad_norm": 0.015145624987781048, + "learning_rate": 0.0006, + "loss": 4.267950534820557, + "step": 4861 + }, + { + "epoch": 67.53100436681223, + "grad_norm": 0.014610803686082363, + "learning_rate": 0.0006, + "loss": 4.247499942779541, + "step": 4862 + }, + { + "epoch": 67.54497816593886, + "grad_norm": 0.014860909432172775, + "learning_rate": 0.0006, + "loss": 4.263200759887695, + "step": 4863 + }, + { + "epoch": 67.5589519650655, + "grad_norm": 0.01718984916806221, + "learning_rate": 0.0006, + "loss": 4.290019989013672, + "step": 4864 + }, + { + "epoch": 67.57292576419214, + "grad_norm": 0.01558963768184185, + "learning_rate": 0.0006, + "loss": 4.223816871643066, + "step": 4865 + }, + { + "epoch": 67.58689956331878, + "grad_norm": 0.014927188865840435, + "learning_rate": 0.0006, + "loss": 4.363586902618408, + "step": 4866 + }, + { + "epoch": 67.60087336244541, + "grad_norm": 0.01573595404624939, + "learning_rate": 0.0006, + "loss": 4.254765510559082, + "step": 4867 + }, + { + "epoch": 67.61484716157206, + "grad_norm": 0.015618929639458656, + "learning_rate": 0.0006, + "loss": 4.25072717666626, + "step": 4868 + }, + { + "epoch": 67.62882096069869, + "grad_norm": 0.01605590060353279, + "learning_rate": 0.0006, + "loss": 4.25208854675293, + "step": 4869 + }, + { + "epoch": 67.64279475982532, + "grad_norm": 0.01804601214826107, + "learning_rate": 0.0006, + "loss": 4.351354122161865, + "step": 4870 + }, + { + "epoch": 67.65676855895197, + "grad_norm": 0.020595047622919083, + "learning_rate": 0.0006, + "loss": 4.193732738494873, + "step": 4871 + }, + { + "epoch": 67.6707423580786, + "grad_norm": 0.023037966340780258, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 4872 + }, + { + "epoch": 67.68471615720524, + "grad_norm": 0.020680895075201988, + "learning_rate": 0.0006, + "loss": 4.314730167388916, + "step": 4873 + }, + { + "epoch": 67.69868995633188, + "grad_norm": 0.01887434534728527, + "learning_rate": 0.0006, + "loss": 4.252305030822754, + "step": 4874 + }, + { + "epoch": 67.71266375545852, + "grad_norm": 0.016333596780896187, + "learning_rate": 0.0006, + "loss": 4.146123886108398, + "step": 4875 + }, + { + "epoch": 67.72663755458515, + "grad_norm": 0.015688810497522354, + "learning_rate": 0.0006, + "loss": 4.2814154624938965, + "step": 4876 + }, + { + "epoch": 67.74061135371178, + "grad_norm": 0.016270726919174194, + "learning_rate": 0.0006, + "loss": 4.323200225830078, + "step": 4877 + }, + { + "epoch": 67.75458515283843, + "grad_norm": 0.016824988648295403, + "learning_rate": 0.0006, + "loss": 4.269501686096191, + "step": 4878 + }, + { + "epoch": 67.76855895196506, + "grad_norm": 0.01730562001466751, + "learning_rate": 0.0006, + "loss": 4.187993049621582, + "step": 4879 + }, + { + "epoch": 67.78253275109171, + "grad_norm": 0.01546655036509037, + "learning_rate": 0.0006, + "loss": 4.340415000915527, + "step": 4880 + }, + { + "epoch": 67.79650655021834, + "grad_norm": 0.014173958450555801, + "learning_rate": 0.0006, + "loss": 4.217419624328613, + "step": 4881 + }, + { + "epoch": 67.81048034934499, + "grad_norm": 0.016810474917292595, + "learning_rate": 0.0006, + "loss": 4.352302074432373, + "step": 4882 + }, + { + "epoch": 67.82445414847162, + "grad_norm": 0.0175373163074255, + "learning_rate": 0.0006, + "loss": 4.20377254486084, + "step": 4883 + }, + { + "epoch": 67.83842794759825, + "grad_norm": 0.015734676271677017, + "learning_rate": 0.0006, + "loss": 4.147495746612549, + "step": 4884 + }, + { + "epoch": 67.8524017467249, + "grad_norm": 0.015091817826032639, + "learning_rate": 0.0006, + "loss": 4.272252082824707, + "step": 4885 + }, + { + "epoch": 67.86637554585153, + "grad_norm": 0.015000219456851482, + "learning_rate": 0.0006, + "loss": 4.302773952484131, + "step": 4886 + }, + { + "epoch": 67.88034934497817, + "grad_norm": 0.013827184215188026, + "learning_rate": 0.0006, + "loss": 4.337242126464844, + "step": 4887 + }, + { + "epoch": 67.8943231441048, + "grad_norm": 0.015308346599340439, + "learning_rate": 0.0006, + "loss": 4.140895366668701, + "step": 4888 + }, + { + "epoch": 67.90829694323143, + "grad_norm": 0.015356178395450115, + "learning_rate": 0.0006, + "loss": 4.312541484832764, + "step": 4889 + }, + { + "epoch": 67.92227074235808, + "grad_norm": 0.0137353939935565, + "learning_rate": 0.0006, + "loss": 4.173184394836426, + "step": 4890 + }, + { + "epoch": 67.93624454148471, + "grad_norm": 0.01332368515431881, + "learning_rate": 0.0006, + "loss": 4.191199779510498, + "step": 4891 + }, + { + "epoch": 67.95021834061136, + "grad_norm": 0.01488130260258913, + "learning_rate": 0.0006, + "loss": 4.331752300262451, + "step": 4892 + }, + { + "epoch": 67.96419213973799, + "grad_norm": 0.015449733473360538, + "learning_rate": 0.0006, + "loss": 4.2385454177856445, + "step": 4893 + }, + { + "epoch": 67.97816593886463, + "grad_norm": 0.01556103490293026, + "learning_rate": 0.0006, + "loss": 4.262353897094727, + "step": 4894 + }, + { + "epoch": 67.99213973799127, + "grad_norm": 0.014613849110901356, + "learning_rate": 0.0006, + "loss": 4.223945617675781, + "step": 4895 + }, + { + "epoch": 68.0, + "grad_norm": 0.015323462896049023, + "learning_rate": 0.0006, + "loss": 4.317168712615967, + "step": 4896 + }, + { + "epoch": 68.0, + "eval_loss": 4.671603679656982, + "eval_runtime": 56.7317, + "eval_samples_per_second": 43.045, + "eval_steps_per_second": 1.357, + "step": 4896 + }, + { + "epoch": 68.01397379912663, + "grad_norm": 0.01910698413848877, + "learning_rate": 0.0006, + "loss": 4.206167221069336, + "step": 4897 + }, + { + "epoch": 68.02794759825328, + "grad_norm": 0.025112485513091087, + "learning_rate": 0.0006, + "loss": 4.384706974029541, + "step": 4898 + }, + { + "epoch": 68.04192139737991, + "grad_norm": 0.02618720941245556, + "learning_rate": 0.0006, + "loss": 4.321046829223633, + "step": 4899 + }, + { + "epoch": 68.05589519650655, + "grad_norm": 0.020853538066148758, + "learning_rate": 0.0006, + "loss": 4.217638969421387, + "step": 4900 + }, + { + "epoch": 68.06986899563319, + "grad_norm": 0.020025717094540596, + "learning_rate": 0.0006, + "loss": 4.306824684143066, + "step": 4901 + }, + { + "epoch": 68.08384279475983, + "grad_norm": 0.023868650197982788, + "learning_rate": 0.0006, + "loss": 4.327455997467041, + "step": 4902 + }, + { + "epoch": 68.09781659388646, + "grad_norm": 0.022279081866145134, + "learning_rate": 0.0006, + "loss": 4.400699615478516, + "step": 4903 + }, + { + "epoch": 68.1117903930131, + "grad_norm": 0.018965687602758408, + "learning_rate": 0.0006, + "loss": 4.19097900390625, + "step": 4904 + }, + { + "epoch": 68.12576419213974, + "grad_norm": 0.01936378888785839, + "learning_rate": 0.0006, + "loss": 4.206711769104004, + "step": 4905 + }, + { + "epoch": 68.13973799126637, + "grad_norm": 0.01811872608959675, + "learning_rate": 0.0006, + "loss": 4.204622268676758, + "step": 4906 + }, + { + "epoch": 68.15371179039302, + "grad_norm": 0.018530402332544327, + "learning_rate": 0.0006, + "loss": 4.289602279663086, + "step": 4907 + }, + { + "epoch": 68.16768558951965, + "grad_norm": 0.01755962148308754, + "learning_rate": 0.0006, + "loss": 4.317145347595215, + "step": 4908 + }, + { + "epoch": 68.18165938864628, + "grad_norm": 0.020413702353835106, + "learning_rate": 0.0006, + "loss": 4.2882843017578125, + "step": 4909 + }, + { + "epoch": 68.19563318777293, + "grad_norm": 0.020064299926161766, + "learning_rate": 0.0006, + "loss": 4.248260021209717, + "step": 4910 + }, + { + "epoch": 68.20960698689956, + "grad_norm": 0.017415888607501984, + "learning_rate": 0.0006, + "loss": 4.370295524597168, + "step": 4911 + }, + { + "epoch": 68.2235807860262, + "grad_norm": 0.015005858615040779, + "learning_rate": 0.0006, + "loss": 4.319567680358887, + "step": 4912 + }, + { + "epoch": 68.23755458515284, + "grad_norm": 0.016349777579307556, + "learning_rate": 0.0006, + "loss": 4.301776885986328, + "step": 4913 + }, + { + "epoch": 68.25152838427948, + "grad_norm": 0.015779560431838036, + "learning_rate": 0.0006, + "loss": 4.271171569824219, + "step": 4914 + }, + { + "epoch": 68.26550218340611, + "grad_norm": 0.013237644918262959, + "learning_rate": 0.0006, + "loss": 4.324908256530762, + "step": 4915 + }, + { + "epoch": 68.27947598253274, + "grad_norm": 0.014076225459575653, + "learning_rate": 0.0006, + "loss": 4.177175998687744, + "step": 4916 + }, + { + "epoch": 68.29344978165939, + "grad_norm": 0.013906744308769703, + "learning_rate": 0.0006, + "loss": 4.11588191986084, + "step": 4917 + }, + { + "epoch": 68.30742358078602, + "grad_norm": 0.013939064927399158, + "learning_rate": 0.0006, + "loss": 4.183963775634766, + "step": 4918 + }, + { + "epoch": 68.32139737991267, + "grad_norm": 0.014572170563042164, + "learning_rate": 0.0006, + "loss": 4.19586181640625, + "step": 4919 + }, + { + "epoch": 68.3353711790393, + "grad_norm": 0.014085350558161736, + "learning_rate": 0.0006, + "loss": 4.170892238616943, + "step": 4920 + }, + { + "epoch": 68.34934497816595, + "grad_norm": 0.014357863925397396, + "learning_rate": 0.0006, + "loss": 4.321155548095703, + "step": 4921 + }, + { + "epoch": 68.36331877729258, + "grad_norm": 0.01461675763130188, + "learning_rate": 0.0006, + "loss": 4.293334007263184, + "step": 4922 + }, + { + "epoch": 68.37729257641921, + "grad_norm": 0.014828486368060112, + "learning_rate": 0.0006, + "loss": 4.1439714431762695, + "step": 4923 + }, + { + "epoch": 68.39126637554585, + "grad_norm": 0.013730363920331001, + "learning_rate": 0.0006, + "loss": 4.287973880767822, + "step": 4924 + }, + { + "epoch": 68.40524017467249, + "grad_norm": 0.013852203264832497, + "learning_rate": 0.0006, + "loss": 4.248824119567871, + "step": 4925 + }, + { + "epoch": 68.41921397379913, + "grad_norm": 0.014005862176418304, + "learning_rate": 0.0006, + "loss": 4.293768882751465, + "step": 4926 + }, + { + "epoch": 68.43318777292576, + "grad_norm": 0.014435851015150547, + "learning_rate": 0.0006, + "loss": 4.204813003540039, + "step": 4927 + }, + { + "epoch": 68.44716157205241, + "grad_norm": 0.015305934473872185, + "learning_rate": 0.0006, + "loss": 4.329364776611328, + "step": 4928 + }, + { + "epoch": 68.46113537117904, + "grad_norm": 0.014356814324855804, + "learning_rate": 0.0006, + "loss": 4.264206409454346, + "step": 4929 + }, + { + "epoch": 68.47510917030567, + "grad_norm": 0.01432347297668457, + "learning_rate": 0.0006, + "loss": 4.29440450668335, + "step": 4930 + }, + { + "epoch": 68.48908296943232, + "grad_norm": 0.012672432698309422, + "learning_rate": 0.0006, + "loss": 4.258024215698242, + "step": 4931 + }, + { + "epoch": 68.50305676855895, + "grad_norm": 0.014209977351129055, + "learning_rate": 0.0006, + "loss": 4.216668128967285, + "step": 4932 + }, + { + "epoch": 68.5170305676856, + "grad_norm": 0.014606663957238197, + "learning_rate": 0.0006, + "loss": 4.294184684753418, + "step": 4933 + }, + { + "epoch": 68.53100436681223, + "grad_norm": 0.014986990951001644, + "learning_rate": 0.0006, + "loss": 4.382591247558594, + "step": 4934 + }, + { + "epoch": 68.54497816593886, + "grad_norm": 0.013998783193528652, + "learning_rate": 0.0006, + "loss": 4.195143222808838, + "step": 4935 + }, + { + "epoch": 68.5589519650655, + "grad_norm": 0.015011102892458439, + "learning_rate": 0.0006, + "loss": 4.313785076141357, + "step": 4936 + }, + { + "epoch": 68.57292576419214, + "grad_norm": 0.01641209051012993, + "learning_rate": 0.0006, + "loss": 4.322803020477295, + "step": 4937 + }, + { + "epoch": 68.58689956331878, + "grad_norm": 0.016207464039325714, + "learning_rate": 0.0006, + "loss": 4.311431884765625, + "step": 4938 + }, + { + "epoch": 68.60087336244541, + "grad_norm": 0.01747811958193779, + "learning_rate": 0.0006, + "loss": 4.310845375061035, + "step": 4939 + }, + { + "epoch": 68.61484716157206, + "grad_norm": 0.019475717097520828, + "learning_rate": 0.0006, + "loss": 4.338624954223633, + "step": 4940 + }, + { + "epoch": 68.62882096069869, + "grad_norm": 0.018641719594597816, + "learning_rate": 0.0006, + "loss": 4.34913969039917, + "step": 4941 + }, + { + "epoch": 68.64279475982532, + "grad_norm": 0.018376432359218597, + "learning_rate": 0.0006, + "loss": 4.28464412689209, + "step": 4942 + }, + { + "epoch": 68.65676855895197, + "grad_norm": 0.017095010727643967, + "learning_rate": 0.0006, + "loss": 4.217957496643066, + "step": 4943 + }, + { + "epoch": 68.6707423580786, + "grad_norm": 0.016321983188390732, + "learning_rate": 0.0006, + "loss": 4.220042705535889, + "step": 4944 + }, + { + "epoch": 68.68471615720524, + "grad_norm": 0.017128000035881996, + "learning_rate": 0.0006, + "loss": 4.293500900268555, + "step": 4945 + }, + { + "epoch": 68.69868995633188, + "grad_norm": 0.015592411160469055, + "learning_rate": 0.0006, + "loss": 4.195798873901367, + "step": 4946 + }, + { + "epoch": 68.71266375545852, + "grad_norm": 0.016607079654932022, + "learning_rate": 0.0006, + "loss": 4.2654337882995605, + "step": 4947 + }, + { + "epoch": 68.72663755458515, + "grad_norm": 0.01545762736350298, + "learning_rate": 0.0006, + "loss": 4.303828239440918, + "step": 4948 + }, + { + "epoch": 68.74061135371178, + "grad_norm": 0.014328422956168652, + "learning_rate": 0.0006, + "loss": 4.29379940032959, + "step": 4949 + }, + { + "epoch": 68.75458515283843, + "grad_norm": 0.016516804695129395, + "learning_rate": 0.0006, + "loss": 4.2317214012146, + "step": 4950 + }, + { + "epoch": 68.76855895196506, + "grad_norm": 0.016324171796441078, + "learning_rate": 0.0006, + "loss": 4.278043746948242, + "step": 4951 + }, + { + "epoch": 68.78253275109171, + "grad_norm": 0.015511159785091877, + "learning_rate": 0.0006, + "loss": 4.188244342803955, + "step": 4952 + }, + { + "epoch": 68.79650655021834, + "grad_norm": 0.01676103286445141, + "learning_rate": 0.0006, + "loss": 4.2169976234436035, + "step": 4953 + }, + { + "epoch": 68.81048034934499, + "grad_norm": 0.015428312122821808, + "learning_rate": 0.0006, + "loss": 4.4004716873168945, + "step": 4954 + }, + { + "epoch": 68.82445414847162, + "grad_norm": 0.0162410456687212, + "learning_rate": 0.0006, + "loss": 4.282752513885498, + "step": 4955 + }, + { + "epoch": 68.83842794759825, + "grad_norm": 0.017578862607479095, + "learning_rate": 0.0006, + "loss": 4.098004341125488, + "step": 4956 + }, + { + "epoch": 68.8524017467249, + "grad_norm": 0.01805749163031578, + "learning_rate": 0.0006, + "loss": 4.277988433837891, + "step": 4957 + }, + { + "epoch": 68.86637554585153, + "grad_norm": 0.016728045418858528, + "learning_rate": 0.0006, + "loss": 4.29766321182251, + "step": 4958 + }, + { + "epoch": 68.88034934497817, + "grad_norm": 0.014899657107889652, + "learning_rate": 0.0006, + "loss": 4.329736709594727, + "step": 4959 + }, + { + "epoch": 68.8943231441048, + "grad_norm": 0.013441305607557297, + "learning_rate": 0.0006, + "loss": 4.2873215675354, + "step": 4960 + }, + { + "epoch": 68.90829694323143, + "grad_norm": 0.01554128248244524, + "learning_rate": 0.0006, + "loss": 4.278327941894531, + "step": 4961 + }, + { + "epoch": 68.92227074235808, + "grad_norm": 0.01806696318089962, + "learning_rate": 0.0006, + "loss": 4.38275146484375, + "step": 4962 + }, + { + "epoch": 68.93624454148471, + "grad_norm": 0.017025409266352654, + "learning_rate": 0.0006, + "loss": 4.210447311401367, + "step": 4963 + }, + { + "epoch": 68.95021834061136, + "grad_norm": 0.01597420684993267, + "learning_rate": 0.0006, + "loss": 4.309865951538086, + "step": 4964 + }, + { + "epoch": 68.96419213973799, + "grad_norm": 0.015835624188184738, + "learning_rate": 0.0006, + "loss": 4.225904941558838, + "step": 4965 + }, + { + "epoch": 68.97816593886463, + "grad_norm": 0.018170492723584175, + "learning_rate": 0.0006, + "loss": 4.248478889465332, + "step": 4966 + }, + { + "epoch": 68.99213973799127, + "grad_norm": 0.017372991889715195, + "learning_rate": 0.0006, + "loss": 4.283486843109131, + "step": 4967 + }, + { + "epoch": 69.0, + "grad_norm": 0.016305390745401382, + "learning_rate": 0.0006, + "loss": 4.273636817932129, + "step": 4968 + }, + { + "epoch": 69.0, + "eval_loss": 4.66873836517334, + "eval_runtime": 57.0791, + "eval_samples_per_second": 42.783, + "eval_steps_per_second": 1.349, + "step": 4968 + }, + { + "epoch": 69.01397379912663, + "grad_norm": 0.016567355021834373, + "learning_rate": 0.0006, + "loss": 4.269988059997559, + "step": 4969 + }, + { + "epoch": 69.02794759825328, + "grad_norm": 0.017476852983236313, + "learning_rate": 0.0006, + "loss": 4.2518720626831055, + "step": 4970 + }, + { + "epoch": 69.04192139737991, + "grad_norm": 0.017070338129997253, + "learning_rate": 0.0006, + "loss": 4.238024711608887, + "step": 4971 + }, + { + "epoch": 69.05589519650655, + "grad_norm": 0.01639373227953911, + "learning_rate": 0.0006, + "loss": 4.188558101654053, + "step": 4972 + }, + { + "epoch": 69.06986899563319, + "grad_norm": 0.01856113411486149, + "learning_rate": 0.0006, + "loss": 4.339556694030762, + "step": 4973 + }, + { + "epoch": 69.08384279475983, + "grad_norm": 0.01762578636407852, + "learning_rate": 0.0006, + "loss": 4.141518592834473, + "step": 4974 + }, + { + "epoch": 69.09781659388646, + "grad_norm": 0.015437949448823929, + "learning_rate": 0.0006, + "loss": 4.225461959838867, + "step": 4975 + }, + { + "epoch": 69.1117903930131, + "grad_norm": 0.015293503180146217, + "learning_rate": 0.0006, + "loss": 4.290884971618652, + "step": 4976 + }, + { + "epoch": 69.12576419213974, + "grad_norm": 0.016259880736470222, + "learning_rate": 0.0006, + "loss": 4.121919631958008, + "step": 4977 + }, + { + "epoch": 69.13973799126637, + "grad_norm": 0.017225584015250206, + "learning_rate": 0.0006, + "loss": 4.136716842651367, + "step": 4978 + }, + { + "epoch": 69.15371179039302, + "grad_norm": 0.01699289306998253, + "learning_rate": 0.0006, + "loss": 4.193601608276367, + "step": 4979 + }, + { + "epoch": 69.16768558951965, + "grad_norm": 0.015613161027431488, + "learning_rate": 0.0006, + "loss": 4.288212299346924, + "step": 4980 + }, + { + "epoch": 69.18165938864628, + "grad_norm": 0.01643933728337288, + "learning_rate": 0.0006, + "loss": 4.324707984924316, + "step": 4981 + }, + { + "epoch": 69.19563318777293, + "grad_norm": 0.01648905873298645, + "learning_rate": 0.0006, + "loss": 4.287806510925293, + "step": 4982 + }, + { + "epoch": 69.20960698689956, + "grad_norm": 0.015444950200617313, + "learning_rate": 0.0006, + "loss": 4.200622081756592, + "step": 4983 + }, + { + "epoch": 69.2235807860262, + "grad_norm": 0.016034258529543877, + "learning_rate": 0.0006, + "loss": 4.2193827629089355, + "step": 4984 + }, + { + "epoch": 69.23755458515284, + "grad_norm": 0.017984388396143913, + "learning_rate": 0.0006, + "loss": 4.18237829208374, + "step": 4985 + }, + { + "epoch": 69.25152838427948, + "grad_norm": 0.0195505041629076, + "learning_rate": 0.0006, + "loss": 4.216727256774902, + "step": 4986 + }, + { + "epoch": 69.26550218340611, + "grad_norm": 0.017779147252440453, + "learning_rate": 0.0006, + "loss": 4.278628826141357, + "step": 4987 + }, + { + "epoch": 69.27947598253274, + "grad_norm": 0.017119476571679115, + "learning_rate": 0.0006, + "loss": 4.2409138679504395, + "step": 4988 + }, + { + "epoch": 69.29344978165939, + "grad_norm": 0.016840364784002304, + "learning_rate": 0.0006, + "loss": 4.207575798034668, + "step": 4989 + }, + { + "epoch": 69.30742358078602, + "grad_norm": 0.014593573287129402, + "learning_rate": 0.0006, + "loss": 4.164905548095703, + "step": 4990 + }, + { + "epoch": 69.32139737991267, + "grad_norm": 0.013922630809247494, + "learning_rate": 0.0006, + "loss": 4.2696990966796875, + "step": 4991 + }, + { + "epoch": 69.3353711790393, + "grad_norm": 0.01577564887702465, + "learning_rate": 0.0006, + "loss": 4.201589107513428, + "step": 4992 + }, + { + "epoch": 69.34934497816595, + "grad_norm": 0.016392096877098083, + "learning_rate": 0.0006, + "loss": 4.248831748962402, + "step": 4993 + }, + { + "epoch": 69.36331877729258, + "grad_norm": 0.015961311757564545, + "learning_rate": 0.0006, + "loss": 4.181260108947754, + "step": 4994 + }, + { + "epoch": 69.37729257641921, + "grad_norm": 0.015350551344454288, + "learning_rate": 0.0006, + "loss": 4.230352878570557, + "step": 4995 + }, + { + "epoch": 69.39126637554585, + "grad_norm": 0.015870466828346252, + "learning_rate": 0.0006, + "loss": 4.298305511474609, + "step": 4996 + }, + { + "epoch": 69.40524017467249, + "grad_norm": 0.013716642744839191, + "learning_rate": 0.0006, + "loss": 4.220139503479004, + "step": 4997 + }, + { + "epoch": 69.41921397379913, + "grad_norm": 0.014431777410209179, + "learning_rate": 0.0006, + "loss": 4.288130760192871, + "step": 4998 + }, + { + "epoch": 69.43318777292576, + "grad_norm": 0.01426647324115038, + "learning_rate": 0.0006, + "loss": 4.295767784118652, + "step": 4999 + }, + { + "epoch": 69.44716157205241, + "grad_norm": 0.014286615885794163, + "learning_rate": 0.0006, + "loss": 4.38106632232666, + "step": 5000 + }, + { + "epoch": 69.46113537117904, + "grad_norm": 0.017990585416555405, + "learning_rate": 0.0006, + "loss": 4.1593017578125, + "step": 5001 + }, + { + "epoch": 69.47510917030567, + "grad_norm": 0.018846049904823303, + "learning_rate": 0.0006, + "loss": 4.216091632843018, + "step": 5002 + }, + { + "epoch": 69.48908296943232, + "grad_norm": 0.018179181963205338, + "learning_rate": 0.0006, + "loss": 4.240332126617432, + "step": 5003 + }, + { + "epoch": 69.50305676855895, + "grad_norm": 0.016693448647856712, + "learning_rate": 0.0006, + "loss": 4.256624221801758, + "step": 5004 + }, + { + "epoch": 69.5170305676856, + "grad_norm": 0.01718989387154579, + "learning_rate": 0.0006, + "loss": 4.300940990447998, + "step": 5005 + }, + { + "epoch": 69.53100436681223, + "grad_norm": 0.020236829295754433, + "learning_rate": 0.0006, + "loss": 4.289146423339844, + "step": 5006 + }, + { + "epoch": 69.54497816593886, + "grad_norm": 0.024860285222530365, + "learning_rate": 0.0006, + "loss": 4.216216087341309, + "step": 5007 + }, + { + "epoch": 69.5589519650655, + "grad_norm": 0.02559969387948513, + "learning_rate": 0.0006, + "loss": 4.262890815734863, + "step": 5008 + }, + { + "epoch": 69.57292576419214, + "grad_norm": 0.019180957227945328, + "learning_rate": 0.0006, + "loss": 4.230309009552002, + "step": 5009 + }, + { + "epoch": 69.58689956331878, + "grad_norm": 0.020442591980099678, + "learning_rate": 0.0006, + "loss": 4.291759490966797, + "step": 5010 + }, + { + "epoch": 69.60087336244541, + "grad_norm": 0.0212619137018919, + "learning_rate": 0.0006, + "loss": 4.3529839515686035, + "step": 5011 + }, + { + "epoch": 69.61484716157206, + "grad_norm": 0.02067175693809986, + "learning_rate": 0.0006, + "loss": 4.33197021484375, + "step": 5012 + }, + { + "epoch": 69.62882096069869, + "grad_norm": 0.01802872307598591, + "learning_rate": 0.0006, + "loss": 4.2101969718933105, + "step": 5013 + }, + { + "epoch": 69.64279475982532, + "grad_norm": 0.01787702925503254, + "learning_rate": 0.0006, + "loss": 4.270336627960205, + "step": 5014 + }, + { + "epoch": 69.65676855895197, + "grad_norm": 0.01586683839559555, + "learning_rate": 0.0006, + "loss": 4.222309589385986, + "step": 5015 + }, + { + "epoch": 69.6707423580786, + "grad_norm": 0.01717999391257763, + "learning_rate": 0.0006, + "loss": 4.253704071044922, + "step": 5016 + }, + { + "epoch": 69.68471615720524, + "grad_norm": 0.01594279147684574, + "learning_rate": 0.0006, + "loss": 4.344658851623535, + "step": 5017 + }, + { + "epoch": 69.69868995633188, + "grad_norm": 0.014419942162930965, + "learning_rate": 0.0006, + "loss": 4.2144880294799805, + "step": 5018 + }, + { + "epoch": 69.71266375545852, + "grad_norm": 0.015824224799871445, + "learning_rate": 0.0006, + "loss": 4.226252555847168, + "step": 5019 + }, + { + "epoch": 69.72663755458515, + "grad_norm": 0.013329979032278061, + "learning_rate": 0.0006, + "loss": 4.238587379455566, + "step": 5020 + }, + { + "epoch": 69.74061135371178, + "grad_norm": 0.013745979405939579, + "learning_rate": 0.0006, + "loss": 4.317209243774414, + "step": 5021 + }, + { + "epoch": 69.75458515283843, + "grad_norm": 0.014064167626202106, + "learning_rate": 0.0006, + "loss": 4.316982746124268, + "step": 5022 + }, + { + "epoch": 69.76855895196506, + "grad_norm": 0.014611301012337208, + "learning_rate": 0.0006, + "loss": 4.212728500366211, + "step": 5023 + }, + { + "epoch": 69.78253275109171, + "grad_norm": 0.014947416260838509, + "learning_rate": 0.0006, + "loss": 4.319550514221191, + "step": 5024 + }, + { + "epoch": 69.79650655021834, + "grad_norm": 0.014780526980757713, + "learning_rate": 0.0006, + "loss": 4.1888933181762695, + "step": 5025 + }, + { + "epoch": 69.81048034934499, + "grad_norm": 0.01611459068953991, + "learning_rate": 0.0006, + "loss": 4.239692211151123, + "step": 5026 + }, + { + "epoch": 69.82445414847162, + "grad_norm": 0.015395848080515862, + "learning_rate": 0.0006, + "loss": 4.308794021606445, + "step": 5027 + }, + { + "epoch": 69.83842794759825, + "grad_norm": 0.016113396733999252, + "learning_rate": 0.0006, + "loss": 4.328035354614258, + "step": 5028 + }, + { + "epoch": 69.8524017467249, + "grad_norm": 0.016386888921260834, + "learning_rate": 0.0006, + "loss": 4.272579193115234, + "step": 5029 + }, + { + "epoch": 69.86637554585153, + "grad_norm": 0.013675205409526825, + "learning_rate": 0.0006, + "loss": 4.226552963256836, + "step": 5030 + }, + { + "epoch": 69.88034934497817, + "grad_norm": 0.01472668256610632, + "learning_rate": 0.0006, + "loss": 4.224167346954346, + "step": 5031 + }, + { + "epoch": 69.8943231441048, + "grad_norm": 0.014958228915929794, + "learning_rate": 0.0006, + "loss": 4.232619285583496, + "step": 5032 + }, + { + "epoch": 69.90829694323143, + "grad_norm": 0.012926832772791386, + "learning_rate": 0.0006, + "loss": 4.27107048034668, + "step": 5033 + }, + { + "epoch": 69.92227074235808, + "grad_norm": 0.013138801790773869, + "learning_rate": 0.0006, + "loss": 4.288477420806885, + "step": 5034 + }, + { + "epoch": 69.93624454148471, + "grad_norm": 0.013415461406111717, + "learning_rate": 0.0006, + "loss": 4.2033257484436035, + "step": 5035 + }, + { + "epoch": 69.95021834061136, + "grad_norm": 0.013807233422994614, + "learning_rate": 0.0006, + "loss": 4.274764537811279, + "step": 5036 + }, + { + "epoch": 69.96419213973799, + "grad_norm": 0.013929193839430809, + "learning_rate": 0.0006, + "loss": 4.1854352951049805, + "step": 5037 + }, + { + "epoch": 69.97816593886463, + "grad_norm": 0.015497001819312572, + "learning_rate": 0.0006, + "loss": 4.226498126983643, + "step": 5038 + }, + { + "epoch": 69.99213973799127, + "grad_norm": 0.01874048449099064, + "learning_rate": 0.0006, + "loss": 4.304137706756592, + "step": 5039 + }, + { + "epoch": 70.0, + "grad_norm": 0.021745791658759117, + "learning_rate": 0.0006, + "loss": 4.206809997558594, + "step": 5040 + }, + { + "epoch": 70.0, + "eval_loss": 4.675539970397949, + "eval_runtime": 56.9034, + "eval_samples_per_second": 42.915, + "eval_steps_per_second": 1.353, + "step": 5040 + }, + { + "epoch": 70.01397379912663, + "grad_norm": 0.02124839462339878, + "learning_rate": 0.0006, + "loss": 4.198199272155762, + "step": 5041 + }, + { + "epoch": 70.02794759825328, + "grad_norm": 0.019432688131928444, + "learning_rate": 0.0006, + "loss": 4.187261581420898, + "step": 5042 + }, + { + "epoch": 70.04192139737991, + "grad_norm": 0.01769290305674076, + "learning_rate": 0.0006, + "loss": 4.155284881591797, + "step": 5043 + }, + { + "epoch": 70.05589519650655, + "grad_norm": 0.01751520484685898, + "learning_rate": 0.0006, + "loss": 4.386857986450195, + "step": 5044 + }, + { + "epoch": 70.06986899563319, + "grad_norm": 0.018138699233531952, + "learning_rate": 0.0006, + "loss": 4.117342472076416, + "step": 5045 + }, + { + "epoch": 70.08384279475983, + "grad_norm": 0.016331054270267487, + "learning_rate": 0.0006, + "loss": 4.225295066833496, + "step": 5046 + }, + { + "epoch": 70.09781659388646, + "grad_norm": 0.01658729277551174, + "learning_rate": 0.0006, + "loss": 4.143835067749023, + "step": 5047 + }, + { + "epoch": 70.1117903930131, + "grad_norm": 0.016367698088288307, + "learning_rate": 0.0006, + "loss": 4.1469268798828125, + "step": 5048 + }, + { + "epoch": 70.12576419213974, + "grad_norm": 0.015151629224419594, + "learning_rate": 0.0006, + "loss": 4.117629051208496, + "step": 5049 + }, + { + "epoch": 70.13973799126637, + "grad_norm": 0.016434503719210625, + "learning_rate": 0.0006, + "loss": 4.107848644256592, + "step": 5050 + }, + { + "epoch": 70.15371179039302, + "grad_norm": 0.017837485298514366, + "learning_rate": 0.0006, + "loss": 4.10837459564209, + "step": 5051 + }, + { + "epoch": 70.16768558951965, + "grad_norm": 0.0165961142629385, + "learning_rate": 0.0006, + "loss": 4.169414043426514, + "step": 5052 + }, + { + "epoch": 70.18165938864628, + "grad_norm": 0.0157014112919569, + "learning_rate": 0.0006, + "loss": 4.0629706382751465, + "step": 5053 + }, + { + "epoch": 70.19563318777293, + "grad_norm": 0.01581883803009987, + "learning_rate": 0.0006, + "loss": 4.220012664794922, + "step": 5054 + }, + { + "epoch": 70.20960698689956, + "grad_norm": 0.015285917557775974, + "learning_rate": 0.0006, + "loss": 4.275337219238281, + "step": 5055 + }, + { + "epoch": 70.2235807860262, + "grad_norm": 0.014984549954533577, + "learning_rate": 0.0006, + "loss": 4.284764289855957, + "step": 5056 + }, + { + "epoch": 70.23755458515284, + "grad_norm": 0.015172012150287628, + "learning_rate": 0.0006, + "loss": 4.234498023986816, + "step": 5057 + }, + { + "epoch": 70.25152838427948, + "grad_norm": 0.016846632584929466, + "learning_rate": 0.0006, + "loss": 4.181689262390137, + "step": 5058 + }, + { + "epoch": 70.26550218340611, + "grad_norm": 0.01920711249113083, + "learning_rate": 0.0006, + "loss": 4.235280513763428, + "step": 5059 + }, + { + "epoch": 70.27947598253274, + "grad_norm": 0.017765291035175323, + "learning_rate": 0.0006, + "loss": 4.238955497741699, + "step": 5060 + }, + { + "epoch": 70.29344978165939, + "grad_norm": 0.015436639077961445, + "learning_rate": 0.0006, + "loss": 4.232516288757324, + "step": 5061 + }, + { + "epoch": 70.30742358078602, + "grad_norm": 0.015660086646676064, + "learning_rate": 0.0006, + "loss": 4.260349273681641, + "step": 5062 + }, + { + "epoch": 70.32139737991267, + "grad_norm": 0.016095709055662155, + "learning_rate": 0.0006, + "loss": 4.168505668640137, + "step": 5063 + }, + { + "epoch": 70.3353711790393, + "grad_norm": 0.015001940540969372, + "learning_rate": 0.0006, + "loss": 4.161739349365234, + "step": 5064 + }, + { + "epoch": 70.34934497816595, + "grad_norm": 0.014668774791061878, + "learning_rate": 0.0006, + "loss": 4.181029796600342, + "step": 5065 + }, + { + "epoch": 70.36331877729258, + "grad_norm": 0.013266284950077534, + "learning_rate": 0.0006, + "loss": 4.3052873611450195, + "step": 5066 + }, + { + "epoch": 70.37729257641921, + "grad_norm": 0.015396138653159142, + "learning_rate": 0.0006, + "loss": 4.34413480758667, + "step": 5067 + }, + { + "epoch": 70.39126637554585, + "grad_norm": 0.015511523932218552, + "learning_rate": 0.0006, + "loss": 4.235037803649902, + "step": 5068 + }, + { + "epoch": 70.40524017467249, + "grad_norm": 0.014825738966464996, + "learning_rate": 0.0006, + "loss": 4.262448310852051, + "step": 5069 + }, + { + "epoch": 70.41921397379913, + "grad_norm": 0.015309952199459076, + "learning_rate": 0.0006, + "loss": 4.258711814880371, + "step": 5070 + }, + { + "epoch": 70.43318777292576, + "grad_norm": 0.017614055424928665, + "learning_rate": 0.0006, + "loss": 4.286028861999512, + "step": 5071 + }, + { + "epoch": 70.44716157205241, + "grad_norm": 0.0166039876639843, + "learning_rate": 0.0006, + "loss": 4.149197578430176, + "step": 5072 + }, + { + "epoch": 70.46113537117904, + "grad_norm": 0.016911419108510017, + "learning_rate": 0.0006, + "loss": 4.271981239318848, + "step": 5073 + }, + { + "epoch": 70.47510917030567, + "grad_norm": 0.016806820407509804, + "learning_rate": 0.0006, + "loss": 4.213315010070801, + "step": 5074 + }, + { + "epoch": 70.48908296943232, + "grad_norm": 0.015221742913126945, + "learning_rate": 0.0006, + "loss": 4.171091079711914, + "step": 5075 + }, + { + "epoch": 70.50305676855895, + "grad_norm": 0.014175347983837128, + "learning_rate": 0.0006, + "loss": 4.185309410095215, + "step": 5076 + }, + { + "epoch": 70.5170305676856, + "grad_norm": 0.013897798955440521, + "learning_rate": 0.0006, + "loss": 4.250119209289551, + "step": 5077 + }, + { + "epoch": 70.53100436681223, + "grad_norm": 0.014791003428399563, + "learning_rate": 0.0006, + "loss": 4.2504682540893555, + "step": 5078 + }, + { + "epoch": 70.54497816593886, + "grad_norm": 0.016700128093361855, + "learning_rate": 0.0006, + "loss": 4.235134124755859, + "step": 5079 + }, + { + "epoch": 70.5589519650655, + "grad_norm": 0.016679363325238228, + "learning_rate": 0.0006, + "loss": 4.32028865814209, + "step": 5080 + }, + { + "epoch": 70.57292576419214, + "grad_norm": 0.015117401257157326, + "learning_rate": 0.0006, + "loss": 4.210995674133301, + "step": 5081 + }, + { + "epoch": 70.58689956331878, + "grad_norm": 0.014441263861954212, + "learning_rate": 0.0006, + "loss": 4.27849006652832, + "step": 5082 + }, + { + "epoch": 70.60087336244541, + "grad_norm": 0.01523780357092619, + "learning_rate": 0.0006, + "loss": 4.197360992431641, + "step": 5083 + }, + { + "epoch": 70.61484716157206, + "grad_norm": 0.016201738268136978, + "learning_rate": 0.0006, + "loss": 4.28239107131958, + "step": 5084 + }, + { + "epoch": 70.62882096069869, + "grad_norm": 0.016003292053937912, + "learning_rate": 0.0006, + "loss": 4.209715843200684, + "step": 5085 + }, + { + "epoch": 70.64279475982532, + "grad_norm": 0.01472893450409174, + "learning_rate": 0.0006, + "loss": 4.169078826904297, + "step": 5086 + }, + { + "epoch": 70.65676855895197, + "grad_norm": 0.01604326069355011, + "learning_rate": 0.0006, + "loss": 4.090967178344727, + "step": 5087 + }, + { + "epoch": 70.6707423580786, + "grad_norm": 0.01598356105387211, + "learning_rate": 0.0006, + "loss": 4.16630220413208, + "step": 5088 + }, + { + "epoch": 70.68471615720524, + "grad_norm": 0.016988780349493027, + "learning_rate": 0.0006, + "loss": 4.117822170257568, + "step": 5089 + }, + { + "epoch": 70.69868995633188, + "grad_norm": 0.018582893535494804, + "learning_rate": 0.0006, + "loss": 4.192902088165283, + "step": 5090 + }, + { + "epoch": 70.71266375545852, + "grad_norm": 0.01793155074119568, + "learning_rate": 0.0006, + "loss": 4.314806938171387, + "step": 5091 + }, + { + "epoch": 70.72663755458515, + "grad_norm": 0.016286427155137062, + "learning_rate": 0.0006, + "loss": 4.2137556076049805, + "step": 5092 + }, + { + "epoch": 70.74061135371178, + "grad_norm": 0.014718376100063324, + "learning_rate": 0.0006, + "loss": 4.3073015213012695, + "step": 5093 + }, + { + "epoch": 70.75458515283843, + "grad_norm": 0.014675077050924301, + "learning_rate": 0.0006, + "loss": 4.160961151123047, + "step": 5094 + }, + { + "epoch": 70.76855895196506, + "grad_norm": 0.015550913289189339, + "learning_rate": 0.0006, + "loss": 4.322887420654297, + "step": 5095 + }, + { + "epoch": 70.78253275109171, + "grad_norm": 0.01453235000371933, + "learning_rate": 0.0006, + "loss": 4.2915873527526855, + "step": 5096 + }, + { + "epoch": 70.79650655021834, + "grad_norm": 0.01459164172410965, + "learning_rate": 0.0006, + "loss": 4.247288227081299, + "step": 5097 + }, + { + "epoch": 70.81048034934499, + "grad_norm": 0.015395080670714378, + "learning_rate": 0.0006, + "loss": 4.360955715179443, + "step": 5098 + }, + { + "epoch": 70.82445414847162, + "grad_norm": 0.01317683607339859, + "learning_rate": 0.0006, + "loss": 4.181969165802002, + "step": 5099 + }, + { + "epoch": 70.83842794759825, + "grad_norm": 0.013917932286858559, + "learning_rate": 0.0006, + "loss": 4.264626979827881, + "step": 5100 + }, + { + "epoch": 70.8524017467249, + "grad_norm": 0.015268695540726185, + "learning_rate": 0.0006, + "loss": 4.287857532501221, + "step": 5101 + }, + { + "epoch": 70.86637554585153, + "grad_norm": 0.015531329438090324, + "learning_rate": 0.0006, + "loss": 4.294188499450684, + "step": 5102 + }, + { + "epoch": 70.88034934497817, + "grad_norm": 0.015607129782438278, + "learning_rate": 0.0006, + "loss": 4.331040859222412, + "step": 5103 + }, + { + "epoch": 70.8943231441048, + "grad_norm": 0.015297149308025837, + "learning_rate": 0.0006, + "loss": 4.243650436401367, + "step": 5104 + }, + { + "epoch": 70.90829694323143, + "grad_norm": 0.014790025539696217, + "learning_rate": 0.0006, + "loss": 4.159859657287598, + "step": 5105 + }, + { + "epoch": 70.92227074235808, + "grad_norm": 0.013466274365782738, + "learning_rate": 0.0006, + "loss": 4.221992492675781, + "step": 5106 + }, + { + "epoch": 70.93624454148471, + "grad_norm": 0.01361898984760046, + "learning_rate": 0.0006, + "loss": 4.278751373291016, + "step": 5107 + }, + { + "epoch": 70.95021834061136, + "grad_norm": 0.01325751468539238, + "learning_rate": 0.0006, + "loss": 4.166001319885254, + "step": 5108 + }, + { + "epoch": 70.96419213973799, + "grad_norm": 0.012727087363600731, + "learning_rate": 0.0006, + "loss": 4.219997406005859, + "step": 5109 + }, + { + "epoch": 70.97816593886463, + "grad_norm": 0.013966867700219154, + "learning_rate": 0.0006, + "loss": 4.207491397857666, + "step": 5110 + }, + { + "epoch": 70.99213973799127, + "grad_norm": 0.015053587965667248, + "learning_rate": 0.0006, + "loss": 4.343245983123779, + "step": 5111 + }, + { + "epoch": 71.0, + "grad_norm": 0.017723368480801582, + "learning_rate": 0.0006, + "loss": 4.356122016906738, + "step": 5112 + }, + { + "epoch": 71.0, + "eval_loss": 4.598303318023682, + "eval_runtime": 56.7521, + "eval_samples_per_second": 43.029, + "eval_steps_per_second": 1.357, + "step": 5112 + }, + { + "epoch": 71.01397379912663, + "grad_norm": 0.018996279686689377, + "learning_rate": 0.0006, + "loss": 4.234772205352783, + "step": 5113 + }, + { + "epoch": 71.02794759825328, + "grad_norm": 0.02041555754840374, + "learning_rate": 0.0006, + "loss": 4.275816917419434, + "step": 5114 + }, + { + "epoch": 71.04192139737991, + "grad_norm": 0.021955285221338272, + "learning_rate": 0.0006, + "loss": 4.166038513183594, + "step": 5115 + }, + { + "epoch": 71.05589519650655, + "grad_norm": 0.021308740600943565, + "learning_rate": 0.0006, + "loss": 4.13063907623291, + "step": 5116 + }, + { + "epoch": 71.06986899563319, + "grad_norm": 0.01747054047882557, + "learning_rate": 0.0006, + "loss": 4.200522422790527, + "step": 5117 + }, + { + "epoch": 71.08384279475983, + "grad_norm": 0.018544532358646393, + "learning_rate": 0.0006, + "loss": 4.173915386199951, + "step": 5118 + }, + { + "epoch": 71.09781659388646, + "grad_norm": 0.0182361900806427, + "learning_rate": 0.0006, + "loss": 4.27392578125, + "step": 5119 + }, + { + "epoch": 71.1117903930131, + "grad_norm": 0.017062073573470116, + "learning_rate": 0.0006, + "loss": 4.219576358795166, + "step": 5120 + }, + { + "epoch": 71.12576419213974, + "grad_norm": 0.016289666295051575, + "learning_rate": 0.0006, + "loss": 4.133376598358154, + "step": 5121 + }, + { + "epoch": 71.13973799126637, + "grad_norm": 0.014864642173051834, + "learning_rate": 0.0006, + "loss": 4.168276786804199, + "step": 5122 + }, + { + "epoch": 71.15371179039302, + "grad_norm": 0.015309540554881096, + "learning_rate": 0.0006, + "loss": 4.175103187561035, + "step": 5123 + }, + { + "epoch": 71.16768558951965, + "grad_norm": 0.01556059904396534, + "learning_rate": 0.0006, + "loss": 4.289899826049805, + "step": 5124 + }, + { + "epoch": 71.18165938864628, + "grad_norm": 0.01479957066476345, + "learning_rate": 0.0006, + "loss": 4.341495037078857, + "step": 5125 + }, + { + "epoch": 71.19563318777293, + "grad_norm": 0.014762179926037788, + "learning_rate": 0.0006, + "loss": 4.252333641052246, + "step": 5126 + }, + { + "epoch": 71.20960698689956, + "grad_norm": 0.014897473156452179, + "learning_rate": 0.0006, + "loss": 4.2244181632995605, + "step": 5127 + }, + { + "epoch": 71.2235807860262, + "grad_norm": 0.012990163639187813, + "learning_rate": 0.0006, + "loss": 4.237288475036621, + "step": 5128 + }, + { + "epoch": 71.23755458515284, + "grad_norm": 0.012858416885137558, + "learning_rate": 0.0006, + "loss": 4.265738487243652, + "step": 5129 + }, + { + "epoch": 71.25152838427948, + "grad_norm": 0.013454788364470005, + "learning_rate": 0.0006, + "loss": 4.289131164550781, + "step": 5130 + }, + { + "epoch": 71.26550218340611, + "grad_norm": 0.013717238791286945, + "learning_rate": 0.0006, + "loss": 4.2844014167785645, + "step": 5131 + }, + { + "epoch": 71.27947598253274, + "grad_norm": 0.0135411461815238, + "learning_rate": 0.0006, + "loss": 4.139082908630371, + "step": 5132 + }, + { + "epoch": 71.29344978165939, + "grad_norm": 0.01406062301248312, + "learning_rate": 0.0006, + "loss": 4.226719856262207, + "step": 5133 + }, + { + "epoch": 71.30742358078602, + "grad_norm": 0.014027190394699574, + "learning_rate": 0.0006, + "loss": 4.191699981689453, + "step": 5134 + }, + { + "epoch": 71.32139737991267, + "grad_norm": 0.013809907250106335, + "learning_rate": 0.0006, + "loss": 4.242542266845703, + "step": 5135 + }, + { + "epoch": 71.3353711790393, + "grad_norm": 0.013507804833352566, + "learning_rate": 0.0006, + "loss": 4.068206310272217, + "step": 5136 + }, + { + "epoch": 71.34934497816595, + "grad_norm": 0.01432533748447895, + "learning_rate": 0.0006, + "loss": 4.378222942352295, + "step": 5137 + }, + { + "epoch": 71.36331877729258, + "grad_norm": 0.01753002591431141, + "learning_rate": 0.0006, + "loss": 4.174706935882568, + "step": 5138 + }, + { + "epoch": 71.37729257641921, + "grad_norm": 0.020048344507813454, + "learning_rate": 0.0006, + "loss": 4.19920015335083, + "step": 5139 + }, + { + "epoch": 71.39126637554585, + "grad_norm": 0.018551401793956757, + "learning_rate": 0.0006, + "loss": 4.175182342529297, + "step": 5140 + }, + { + "epoch": 71.40524017467249, + "grad_norm": 0.01598959229886532, + "learning_rate": 0.0006, + "loss": 4.2305169105529785, + "step": 5141 + }, + { + "epoch": 71.41921397379913, + "grad_norm": 0.016029560938477516, + "learning_rate": 0.0006, + "loss": 4.230090618133545, + "step": 5142 + }, + { + "epoch": 71.43318777292576, + "grad_norm": 0.01595657505095005, + "learning_rate": 0.0006, + "loss": 4.138402938842773, + "step": 5143 + }, + { + "epoch": 71.44716157205241, + "grad_norm": 0.01623230241239071, + "learning_rate": 0.0006, + "loss": 4.196106433868408, + "step": 5144 + }, + { + "epoch": 71.46113537117904, + "grad_norm": 0.014057597145438194, + "learning_rate": 0.0006, + "loss": 4.2599897384643555, + "step": 5145 + }, + { + "epoch": 71.47510917030567, + "grad_norm": 0.017161158844828606, + "learning_rate": 0.0006, + "loss": 4.183909893035889, + "step": 5146 + }, + { + "epoch": 71.48908296943232, + "grad_norm": 0.017038581892848015, + "learning_rate": 0.0006, + "loss": 4.373933792114258, + "step": 5147 + }, + { + "epoch": 71.50305676855895, + "grad_norm": 0.015851367264986038, + "learning_rate": 0.0006, + "loss": 4.2620086669921875, + "step": 5148 + }, + { + "epoch": 71.5170305676856, + "grad_norm": 0.014598303474485874, + "learning_rate": 0.0006, + "loss": 4.197151184082031, + "step": 5149 + }, + { + "epoch": 71.53100436681223, + "grad_norm": 0.01339802611619234, + "learning_rate": 0.0006, + "loss": 4.110706329345703, + "step": 5150 + }, + { + "epoch": 71.54497816593886, + "grad_norm": 0.015567775815725327, + "learning_rate": 0.0006, + "loss": 4.1210246086120605, + "step": 5151 + }, + { + "epoch": 71.5589519650655, + "grad_norm": 0.015616007149219513, + "learning_rate": 0.0006, + "loss": 4.129796028137207, + "step": 5152 + }, + { + "epoch": 71.57292576419214, + "grad_norm": 0.014198964461684227, + "learning_rate": 0.0006, + "loss": 4.193232536315918, + "step": 5153 + }, + { + "epoch": 71.58689956331878, + "grad_norm": 0.015226383693516254, + "learning_rate": 0.0006, + "loss": 4.223779678344727, + "step": 5154 + }, + { + "epoch": 71.60087336244541, + "grad_norm": 0.015610109083354473, + "learning_rate": 0.0006, + "loss": 4.223179817199707, + "step": 5155 + }, + { + "epoch": 71.61484716157206, + "grad_norm": 0.016034303233027458, + "learning_rate": 0.0006, + "loss": 4.196593284606934, + "step": 5156 + }, + { + "epoch": 71.62882096069869, + "grad_norm": 0.015648581087589264, + "learning_rate": 0.0006, + "loss": 4.341497421264648, + "step": 5157 + }, + { + "epoch": 71.64279475982532, + "grad_norm": 0.014117540791630745, + "learning_rate": 0.0006, + "loss": 4.194307327270508, + "step": 5158 + }, + { + "epoch": 71.65676855895197, + "grad_norm": 0.015612686984241009, + "learning_rate": 0.0006, + "loss": 4.290152549743652, + "step": 5159 + }, + { + "epoch": 71.6707423580786, + "grad_norm": 0.01722259446978569, + "learning_rate": 0.0006, + "loss": 4.258018493652344, + "step": 5160 + }, + { + "epoch": 71.68471615720524, + "grad_norm": 0.015923351049423218, + "learning_rate": 0.0006, + "loss": 4.21684455871582, + "step": 5161 + }, + { + "epoch": 71.69868995633188, + "grad_norm": 0.013663450255990028, + "learning_rate": 0.0006, + "loss": 4.175601959228516, + "step": 5162 + }, + { + "epoch": 71.71266375545852, + "grad_norm": 0.017075596377253532, + "learning_rate": 0.0006, + "loss": 4.205138683319092, + "step": 5163 + }, + { + "epoch": 71.72663755458515, + "grad_norm": 0.01723487675189972, + "learning_rate": 0.0006, + "loss": 4.08839225769043, + "step": 5164 + }, + { + "epoch": 71.74061135371178, + "grad_norm": 0.017355505377054214, + "learning_rate": 0.0006, + "loss": 4.165130615234375, + "step": 5165 + }, + { + "epoch": 71.75458515283843, + "grad_norm": 0.01868261583149433, + "learning_rate": 0.0006, + "loss": 4.16977071762085, + "step": 5166 + }, + { + "epoch": 71.76855895196506, + "grad_norm": 0.01897321827709675, + "learning_rate": 0.0006, + "loss": 4.173497200012207, + "step": 5167 + }, + { + "epoch": 71.78253275109171, + "grad_norm": 0.02045825682580471, + "learning_rate": 0.0006, + "loss": 4.176192283630371, + "step": 5168 + }, + { + "epoch": 71.79650655021834, + "grad_norm": 0.022895624861121178, + "learning_rate": 0.0006, + "loss": 4.163009166717529, + "step": 5169 + }, + { + "epoch": 71.81048034934499, + "grad_norm": 0.02062244527041912, + "learning_rate": 0.0006, + "loss": 4.156868934631348, + "step": 5170 + }, + { + "epoch": 71.82445414847162, + "grad_norm": 0.0176058579236269, + "learning_rate": 0.0006, + "loss": 4.154374122619629, + "step": 5171 + }, + { + "epoch": 71.83842794759825, + "grad_norm": 0.018049508333206177, + "learning_rate": 0.0006, + "loss": 4.296828269958496, + "step": 5172 + }, + { + "epoch": 71.8524017467249, + "grad_norm": 0.019466130062937737, + "learning_rate": 0.0006, + "loss": 4.19450569152832, + "step": 5173 + }, + { + "epoch": 71.86637554585153, + "grad_norm": 0.01838165521621704, + "learning_rate": 0.0006, + "loss": 4.11812686920166, + "step": 5174 + }, + { + "epoch": 71.88034934497817, + "grad_norm": 0.01608051173388958, + "learning_rate": 0.0006, + "loss": 4.240464210510254, + "step": 5175 + }, + { + "epoch": 71.8943231441048, + "grad_norm": 0.015807967633008957, + "learning_rate": 0.0006, + "loss": 4.275581359863281, + "step": 5176 + }, + { + "epoch": 71.90829694323143, + "grad_norm": 0.017366575077176094, + "learning_rate": 0.0006, + "loss": 4.064750671386719, + "step": 5177 + }, + { + "epoch": 71.92227074235808, + "grad_norm": 0.016241202130913734, + "learning_rate": 0.0006, + "loss": 4.203697681427002, + "step": 5178 + }, + { + "epoch": 71.93624454148471, + "grad_norm": 0.019105732440948486, + "learning_rate": 0.0006, + "loss": 4.140021800994873, + "step": 5179 + }, + { + "epoch": 71.95021834061136, + "grad_norm": 0.018492674455046654, + "learning_rate": 0.0006, + "loss": 4.2534074783325195, + "step": 5180 + }, + { + "epoch": 71.96419213973799, + "grad_norm": 0.019301170483231544, + "learning_rate": 0.0006, + "loss": 4.381392478942871, + "step": 5181 + }, + { + "epoch": 71.97816593886463, + "grad_norm": 0.019571911543607712, + "learning_rate": 0.0006, + "loss": 4.131056785583496, + "step": 5182 + }, + { + "epoch": 71.99213973799127, + "grad_norm": 0.01859748177230358, + "learning_rate": 0.0006, + "loss": 4.278930187225342, + "step": 5183 + }, + { + "epoch": 72.0, + "grad_norm": 0.02017994411289692, + "learning_rate": 0.0006, + "loss": 4.116464614868164, + "step": 5184 + }, + { + "epoch": 72.0, + "eval_loss": 4.60784912109375, + "eval_runtime": 56.3242, + "eval_samples_per_second": 43.356, + "eval_steps_per_second": 1.367, + "step": 5184 + }, + { + "epoch": 72.01397379912663, + "grad_norm": 0.030090296640992165, + "learning_rate": 0.0006, + "loss": 4.218588352203369, + "step": 5185 + }, + { + "epoch": 72.02794759825328, + "grad_norm": 1.330443024635315, + "learning_rate": 0.0006, + "loss": 4.417870044708252, + "step": 5186 + }, + { + "epoch": 72.04192139737991, + "grad_norm": 0.6991518139839172, + "learning_rate": 0.0006, + "loss": 8.414234161376953, + "step": 5187 + }, + { + "epoch": 72.05589519650655, + "grad_norm": 0.2631438374519348, + "learning_rate": 0.0006, + "loss": 7.41729736328125, + "step": 5188 + }, + { + "epoch": 72.06986899563319, + "grad_norm": 0.12391754984855652, + "learning_rate": 0.0006, + "loss": 7.209164619445801, + "step": 5189 + }, + { + "epoch": 72.08384279475983, + "grad_norm": 0.13519170880317688, + "learning_rate": 0.0006, + "loss": 6.943539619445801, + "step": 5190 + }, + { + "epoch": 72.09781659388646, + "grad_norm": 1.258439302444458, + "learning_rate": 0.0006, + "loss": 10.267714500427246, + "step": 5191 + }, + { + "epoch": 72.1117903930131, + "grad_norm": 1.0296908617019653, + "learning_rate": 0.0006, + "loss": 9.51768970489502, + "step": 5192 + }, + { + "epoch": 72.12576419213974, + "grad_norm": 0.12273016571998596, + "learning_rate": 0.0006, + "loss": 7.634477615356445, + "step": 5193 + }, + { + "epoch": 72.13973799126637, + "grad_norm": 0.06170198693871498, + "learning_rate": 0.0006, + "loss": 7.592115879058838, + "step": 5194 + }, + { + "epoch": 72.15371179039302, + "grad_norm": 0.06295108795166016, + "learning_rate": 0.0006, + "loss": 7.576314449310303, + "step": 5195 + }, + { + "epoch": 72.16768558951965, + "grad_norm": 0.0666993111371994, + "learning_rate": 0.0006, + "loss": 7.53862190246582, + "step": 5196 + }, + { + "epoch": 72.18165938864628, + "grad_norm": 0.052215222269296646, + "learning_rate": 0.0006, + "loss": 7.502930641174316, + "step": 5197 + }, + { + "epoch": 72.19563318777293, + "grad_norm": 0.04978961497545242, + "learning_rate": 0.0006, + "loss": 7.471630096435547, + "step": 5198 + }, + { + "epoch": 72.20960698689956, + "grad_norm": 0.07275046408176422, + "learning_rate": 0.0006, + "loss": 7.496924877166748, + "step": 5199 + }, + { + "epoch": 72.2235807860262, + "grad_norm": 0.04491819068789482, + "learning_rate": 0.0006, + "loss": 7.424494743347168, + "step": 5200 + }, + { + "epoch": 72.23755458515284, + "grad_norm": 0.043718695640563965, + "learning_rate": 0.0006, + "loss": 7.441090106964111, + "step": 5201 + }, + { + "epoch": 72.25152838427948, + "grad_norm": 0.04483497887849808, + "learning_rate": 0.0006, + "loss": 7.3753862380981445, + "step": 5202 + }, + { + "epoch": 72.26550218340611, + "grad_norm": 0.03463944420218468, + "learning_rate": 0.0006, + "loss": 7.375782012939453, + "step": 5203 + }, + { + "epoch": 72.27947598253274, + "grad_norm": 0.0397045724093914, + "learning_rate": 0.0006, + "loss": 7.358898639678955, + "step": 5204 + }, + { + "epoch": 72.29344978165939, + "grad_norm": 0.029339438304305077, + "learning_rate": 0.0006, + "loss": 7.3774871826171875, + "step": 5205 + }, + { + "epoch": 72.30742358078602, + "grad_norm": 0.025469699874520302, + "learning_rate": 0.0006, + "loss": 7.324203014373779, + "step": 5206 + }, + { + "epoch": 72.32139737991267, + "grad_norm": 0.030497785657644272, + "learning_rate": 0.0006, + "loss": 7.329351902008057, + "step": 5207 + }, + { + "epoch": 72.3353711790393, + "grad_norm": 0.03382078558206558, + "learning_rate": 0.0006, + "loss": 7.289940357208252, + "step": 5208 + }, + { + "epoch": 72.34934497816595, + "grad_norm": 0.019869405776262283, + "learning_rate": 0.0006, + "loss": 7.272582054138184, + "step": 5209 + }, + { + "epoch": 72.36331877729258, + "grad_norm": 0.02385753206908703, + "learning_rate": 0.0006, + "loss": 7.2654266357421875, + "step": 5210 + }, + { + "epoch": 72.37729257641921, + "grad_norm": 0.02794312685728073, + "learning_rate": 0.0006, + "loss": 7.250753402709961, + "step": 5211 + }, + { + "epoch": 72.39126637554585, + "grad_norm": 0.02445952221751213, + "learning_rate": 0.0006, + "loss": 7.293970108032227, + "step": 5212 + }, + { + "epoch": 72.40524017467249, + "grad_norm": 0.019660986959934235, + "learning_rate": 0.0006, + "loss": 7.1948771476745605, + "step": 5213 + }, + { + "epoch": 72.41921397379913, + "grad_norm": 0.024110861122608185, + "learning_rate": 0.0006, + "loss": 7.224499225616455, + "step": 5214 + }, + { + "epoch": 72.43318777292576, + "grad_norm": 0.021630609408020973, + "learning_rate": 0.0006, + "loss": 7.1736884117126465, + "step": 5215 + }, + { + "epoch": 72.44716157205241, + "grad_norm": 0.02185880020260811, + "learning_rate": 0.0006, + "loss": 7.167229652404785, + "step": 5216 + }, + { + "epoch": 72.46113537117904, + "grad_norm": 0.018913568928837776, + "learning_rate": 0.0006, + "loss": 7.196551322937012, + "step": 5217 + }, + { + "epoch": 72.47510917030567, + "grad_norm": 0.01802486553788185, + "learning_rate": 0.0006, + "loss": 7.144286632537842, + "step": 5218 + }, + { + "epoch": 72.48908296943232, + "grad_norm": 0.020386485382914543, + "learning_rate": 0.0006, + "loss": 7.196260929107666, + "step": 5219 + }, + { + "epoch": 72.50305676855895, + "grad_norm": 0.020077696070075035, + "learning_rate": 0.0006, + "loss": 7.138718605041504, + "step": 5220 + }, + { + "epoch": 72.5170305676856, + "grad_norm": 0.017224183306097984, + "learning_rate": 0.0006, + "loss": 7.119478702545166, + "step": 5221 + }, + { + "epoch": 72.53100436681223, + "grad_norm": 0.020847434177994728, + "learning_rate": 0.0006, + "loss": 7.068922996520996, + "step": 5222 + }, + { + "epoch": 72.54497816593886, + "grad_norm": 0.021764913573861122, + "learning_rate": 0.0006, + "loss": 7.071352958679199, + "step": 5223 + }, + { + "epoch": 72.5589519650655, + "grad_norm": 0.02697215974330902, + "learning_rate": 0.0006, + "loss": 7.033344268798828, + "step": 5224 + }, + { + "epoch": 72.57292576419214, + "grad_norm": 0.042392998933792114, + "learning_rate": 0.0006, + "loss": 7.0589919090271, + "step": 5225 + }, + { + "epoch": 72.58689956331878, + "grad_norm": 0.06616316735744476, + "learning_rate": 0.0006, + "loss": 7.060244083404541, + "step": 5226 + }, + { + "epoch": 72.60087336244541, + "grad_norm": 0.0828910544514656, + "learning_rate": 0.0006, + "loss": 7.057173728942871, + "step": 5227 + }, + { + "epoch": 72.61484716157206, + "grad_norm": 0.046767015010118484, + "learning_rate": 0.0006, + "loss": 7.010563850402832, + "step": 5228 + }, + { + "epoch": 72.62882096069869, + "grad_norm": 0.03831760957837105, + "learning_rate": 0.0006, + "loss": 6.9028825759887695, + "step": 5229 + }, + { + "epoch": 72.64279475982532, + "grad_norm": 0.05109567195177078, + "learning_rate": 0.0006, + "loss": 6.923257827758789, + "step": 5230 + }, + { + "epoch": 72.65676855895197, + "grad_norm": 0.02897934429347515, + "learning_rate": 0.0006, + "loss": 6.823014259338379, + "step": 5231 + }, + { + "epoch": 72.6707423580786, + "grad_norm": 0.03769192099571228, + "learning_rate": 0.0006, + "loss": 6.87416410446167, + "step": 5232 + }, + { + "epoch": 72.68471615720524, + "grad_norm": 0.03677637130022049, + "learning_rate": 0.0006, + "loss": 6.7649030685424805, + "step": 5233 + }, + { + "epoch": 72.69868995633188, + "grad_norm": 0.0371948778629303, + "learning_rate": 0.0006, + "loss": 6.749821662902832, + "step": 5234 + }, + { + "epoch": 72.71266375545852, + "grad_norm": 0.03057795763015747, + "learning_rate": 0.0006, + "loss": 6.748945236206055, + "step": 5235 + }, + { + "epoch": 72.72663755458515, + "grad_norm": 0.028109321370720863, + "learning_rate": 0.0006, + "loss": 6.663548946380615, + "step": 5236 + }, + { + "epoch": 72.74061135371178, + "grad_norm": 0.02442420832812786, + "learning_rate": 0.0006, + "loss": 6.636251449584961, + "step": 5237 + }, + { + "epoch": 72.75458515283843, + "grad_norm": 0.031841423362493515, + "learning_rate": 0.0006, + "loss": 6.606883525848389, + "step": 5238 + }, + { + "epoch": 72.76855895196506, + "grad_norm": 0.028100458905100822, + "learning_rate": 0.0006, + "loss": 6.603745937347412, + "step": 5239 + }, + { + "epoch": 72.78253275109171, + "grad_norm": 0.03222901374101639, + "learning_rate": 0.0006, + "loss": 6.463825225830078, + "step": 5240 + }, + { + "epoch": 72.79650655021834, + "grad_norm": 0.04766521975398064, + "learning_rate": 0.0006, + "loss": 6.43734073638916, + "step": 5241 + }, + { + "epoch": 72.81048034934499, + "grad_norm": 0.09058783948421478, + "learning_rate": 0.0006, + "loss": 6.448520183563232, + "step": 5242 + }, + { + "epoch": 72.82445414847162, + "grad_norm": 0.08995158225297928, + "learning_rate": 0.0006, + "loss": 6.522093296051025, + "step": 5243 + }, + { + "epoch": 72.83842794759825, + "grad_norm": 0.05491040274500847, + "learning_rate": 0.0006, + "loss": 6.37354040145874, + "step": 5244 + }, + { + "epoch": 72.8524017467249, + "grad_norm": 0.05198514088988304, + "learning_rate": 0.0006, + "loss": 6.362032413482666, + "step": 5245 + }, + { + "epoch": 72.86637554585153, + "grad_norm": 0.05464765429496765, + "learning_rate": 0.0006, + "loss": 6.326773643493652, + "step": 5246 + }, + { + "epoch": 72.88034934497817, + "grad_norm": 0.07337361574172974, + "learning_rate": 0.0006, + "loss": 6.297532081604004, + "step": 5247 + }, + { + "epoch": 72.8943231441048, + "grad_norm": 0.10061230510473251, + "learning_rate": 0.0006, + "loss": 6.359557628631592, + "step": 5248 + }, + { + "epoch": 72.90829694323143, + "grad_norm": 0.10005959123373032, + "learning_rate": 0.0006, + "loss": 6.422270774841309, + "step": 5249 + }, + { + "epoch": 72.92227074235808, + "grad_norm": 0.08989301323890686, + "learning_rate": 0.0006, + "loss": 6.289371490478516, + "step": 5250 + }, + { + "epoch": 72.93624454148471, + "grad_norm": 0.06371799111366272, + "learning_rate": 0.0006, + "loss": 6.199618339538574, + "step": 5251 + }, + { + "epoch": 72.95021834061136, + "grad_norm": 0.05932944267988205, + "learning_rate": 0.0006, + "loss": 6.111152648925781, + "step": 5252 + }, + { + "epoch": 72.96419213973799, + "grad_norm": 0.04735502973198891, + "learning_rate": 0.0006, + "loss": 6.026026248931885, + "step": 5253 + }, + { + "epoch": 72.97816593886463, + "grad_norm": 0.04270688816905022, + "learning_rate": 0.0006, + "loss": 6.01606559753418, + "step": 5254 + }, + { + "epoch": 72.99213973799127, + "grad_norm": 0.03774143010377884, + "learning_rate": 0.0006, + "loss": 5.8935441970825195, + "step": 5255 + }, + { + "epoch": 73.0, + "grad_norm": 0.04313631355762482, + "learning_rate": 0.0006, + "loss": 5.842061996459961, + "step": 5256 + }, + { + "epoch": 73.0, + "eval_loss": 5.9985785484313965, + "eval_runtime": 56.3891, + "eval_samples_per_second": 43.306, + "eval_steps_per_second": 1.366, + "step": 5256 + }, + { + "epoch": 73.01397379912663, + "grad_norm": 0.045038655400276184, + "learning_rate": 0.0006, + "loss": 5.8209662437438965, + "step": 5257 + }, + { + "epoch": 73.02794759825328, + "grad_norm": 0.03706910088658333, + "learning_rate": 0.0006, + "loss": 5.806189060211182, + "step": 5258 + }, + { + "epoch": 73.04192139737991, + "grad_norm": 0.04051685333251953, + "learning_rate": 0.0006, + "loss": 5.706869125366211, + "step": 5259 + }, + { + "epoch": 73.05589519650655, + "grad_norm": 0.04849153012037277, + "learning_rate": 0.0006, + "loss": 5.696385383605957, + "step": 5260 + }, + { + "epoch": 73.06986899563319, + "grad_norm": 0.06124403700232506, + "learning_rate": 0.0006, + "loss": 5.618223667144775, + "step": 5261 + }, + { + "epoch": 73.08384279475983, + "grad_norm": 0.07375937700271606, + "learning_rate": 0.0006, + "loss": 5.631045341491699, + "step": 5262 + }, + { + "epoch": 73.09781659388646, + "grad_norm": 0.07413844019174576, + "learning_rate": 0.0006, + "loss": 5.6532182693481445, + "step": 5263 + }, + { + "epoch": 73.1117903930131, + "grad_norm": 0.0633770152926445, + "learning_rate": 0.0006, + "loss": 5.57209587097168, + "step": 5264 + }, + { + "epoch": 73.12576419213974, + "grad_norm": 0.05436766892671585, + "learning_rate": 0.0006, + "loss": 5.53944730758667, + "step": 5265 + }, + { + "epoch": 73.13973799126637, + "grad_norm": 0.04639481008052826, + "learning_rate": 0.0006, + "loss": 5.502948760986328, + "step": 5266 + }, + { + "epoch": 73.15371179039302, + "grad_norm": 0.07556558400392532, + "learning_rate": 0.0006, + "loss": 5.429069995880127, + "step": 5267 + }, + { + "epoch": 73.16768558951965, + "grad_norm": 0.1127929836511612, + "learning_rate": 0.0006, + "loss": 5.545061111450195, + "step": 5268 + }, + { + "epoch": 73.18165938864628, + "grad_norm": 0.07620745897293091, + "learning_rate": 0.0006, + "loss": 5.507102966308594, + "step": 5269 + }, + { + "epoch": 73.19563318777293, + "grad_norm": 0.05076538026332855, + "learning_rate": 0.0006, + "loss": 5.329682350158691, + "step": 5270 + }, + { + "epoch": 73.20960698689956, + "grad_norm": 0.05407547950744629, + "learning_rate": 0.0006, + "loss": 5.3048553466796875, + "step": 5271 + }, + { + "epoch": 73.2235807860262, + "grad_norm": 0.059095583856105804, + "learning_rate": 0.0006, + "loss": 5.292568683624268, + "step": 5272 + }, + { + "epoch": 73.23755458515284, + "grad_norm": 0.06193305924534798, + "learning_rate": 0.0006, + "loss": 5.1870927810668945, + "step": 5273 + }, + { + "epoch": 73.25152838427948, + "grad_norm": 0.062261562794446945, + "learning_rate": 0.0006, + "loss": 5.256819248199463, + "step": 5274 + }, + { + "epoch": 73.26550218340611, + "grad_norm": 0.0732516348361969, + "learning_rate": 0.0006, + "loss": 5.256464004516602, + "step": 5275 + }, + { + "epoch": 73.27947598253274, + "grad_norm": 0.12177979946136475, + "learning_rate": 0.0006, + "loss": 5.284239768981934, + "step": 5276 + }, + { + "epoch": 73.29344978165939, + "grad_norm": 0.10459653288125992, + "learning_rate": 0.0006, + "loss": 5.269713401794434, + "step": 5277 + }, + { + "epoch": 73.30742358078602, + "grad_norm": 0.07854694873094559, + "learning_rate": 0.0006, + "loss": 5.3245391845703125, + "step": 5278 + }, + { + "epoch": 73.32139737991267, + "grad_norm": 0.06030094623565674, + "learning_rate": 0.0006, + "loss": 5.033702850341797, + "step": 5279 + }, + { + "epoch": 73.3353711790393, + "grad_norm": 0.05168461427092552, + "learning_rate": 0.0006, + "loss": 5.140530109405518, + "step": 5280 + }, + { + "epoch": 73.34934497816595, + "grad_norm": 0.041064050048589706, + "learning_rate": 0.0006, + "loss": 5.1085896492004395, + "step": 5281 + }, + { + "epoch": 73.36331877729258, + "grad_norm": 0.03788219019770622, + "learning_rate": 0.0006, + "loss": 5.045009613037109, + "step": 5282 + }, + { + "epoch": 73.37729257641921, + "grad_norm": 0.03851071000099182, + "learning_rate": 0.0006, + "loss": 5.055008888244629, + "step": 5283 + }, + { + "epoch": 73.39126637554585, + "grad_norm": 0.03448539599776268, + "learning_rate": 0.0006, + "loss": 5.10657262802124, + "step": 5284 + }, + { + "epoch": 73.40524017467249, + "grad_norm": 0.026618091389536858, + "learning_rate": 0.0006, + "loss": 5.0342912673950195, + "step": 5285 + }, + { + "epoch": 73.41921397379913, + "grad_norm": 0.03060489147901535, + "learning_rate": 0.0006, + "loss": 5.022700786590576, + "step": 5286 + }, + { + "epoch": 73.43318777292576, + "grad_norm": 0.03362000361084938, + "learning_rate": 0.0006, + "loss": 4.891584396362305, + "step": 5287 + }, + { + "epoch": 73.44716157205241, + "grad_norm": 0.03887505829334259, + "learning_rate": 0.0006, + "loss": 4.772258758544922, + "step": 5288 + }, + { + "epoch": 73.46113537117904, + "grad_norm": 0.07157225161790848, + "learning_rate": 0.0006, + "loss": 4.997142791748047, + "step": 5289 + }, + { + "epoch": 73.47510917030567, + "grad_norm": 0.10180624574422836, + "learning_rate": 0.0006, + "loss": 4.908722877502441, + "step": 5290 + }, + { + "epoch": 73.48908296943232, + "grad_norm": 0.06529122591018677, + "learning_rate": 0.0006, + "loss": 4.98001766204834, + "step": 5291 + }, + { + "epoch": 73.50305676855895, + "grad_norm": 0.055171865969896317, + "learning_rate": 0.0006, + "loss": 4.877438068389893, + "step": 5292 + }, + { + "epoch": 73.5170305676856, + "grad_norm": 0.0318770632147789, + "learning_rate": 0.0006, + "loss": 4.836934566497803, + "step": 5293 + }, + { + "epoch": 73.53100436681223, + "grad_norm": 0.042072564363479614, + "learning_rate": 0.0006, + "loss": 4.919489860534668, + "step": 5294 + }, + { + "epoch": 73.54497816593886, + "grad_norm": 0.029197754338383675, + "learning_rate": 0.0006, + "loss": 4.805579662322998, + "step": 5295 + }, + { + "epoch": 73.5589519650655, + "grad_norm": 0.03249497339129448, + "learning_rate": 0.0006, + "loss": 4.704806327819824, + "step": 5296 + }, + { + "epoch": 73.57292576419214, + "grad_norm": 0.025050945580005646, + "learning_rate": 0.0006, + "loss": 4.936512470245361, + "step": 5297 + }, + { + "epoch": 73.58689956331878, + "grad_norm": 0.029073381796479225, + "learning_rate": 0.0006, + "loss": 4.757592678070068, + "step": 5298 + }, + { + "epoch": 73.60087336244541, + "grad_norm": 0.03465264290571213, + "learning_rate": 0.0006, + "loss": 4.729528427124023, + "step": 5299 + }, + { + "epoch": 73.61484716157206, + "grad_norm": 0.054795749485492706, + "learning_rate": 0.0006, + "loss": 4.77252197265625, + "step": 5300 + }, + { + "epoch": 73.62882096069869, + "grad_norm": 0.18129196763038635, + "learning_rate": 0.0006, + "loss": 4.901700019836426, + "step": 5301 + }, + { + "epoch": 73.64279475982532, + "grad_norm": 0.23274236917495728, + "learning_rate": 0.0006, + "loss": 5.207432270050049, + "step": 5302 + }, + { + "epoch": 73.65676855895197, + "grad_norm": 0.18162772059440613, + "learning_rate": 0.0006, + "loss": 5.326632976531982, + "step": 5303 + }, + { + "epoch": 73.6707423580786, + "grad_norm": 0.11858413368463516, + "learning_rate": 0.0006, + "loss": 5.3730058670043945, + "step": 5304 + }, + { + "epoch": 73.68471615720524, + "grad_norm": 0.06031268090009689, + "learning_rate": 0.0006, + "loss": 5.330936431884766, + "step": 5305 + }, + { + "epoch": 73.69868995633188, + "grad_norm": 0.059304628521203995, + "learning_rate": 0.0006, + "loss": 5.280514717102051, + "step": 5306 + }, + { + "epoch": 73.71266375545852, + "grad_norm": 0.04702131822705269, + "learning_rate": 0.0006, + "loss": 5.288274765014648, + "step": 5307 + }, + { + "epoch": 73.72663755458515, + "grad_norm": 0.037792228162288666, + "learning_rate": 0.0006, + "loss": 5.173289775848389, + "step": 5308 + }, + { + "epoch": 73.74061135371178, + "grad_norm": 0.037763576954603195, + "learning_rate": 0.0006, + "loss": 5.084990978240967, + "step": 5309 + }, + { + "epoch": 73.75458515283843, + "grad_norm": 0.03337346389889717, + "learning_rate": 0.0006, + "loss": 5.121417045593262, + "step": 5310 + }, + { + "epoch": 73.76855895196506, + "grad_norm": 0.03184802085161209, + "learning_rate": 0.0006, + "loss": 5.208918571472168, + "step": 5311 + }, + { + "epoch": 73.78253275109171, + "grad_norm": 0.02832048572599888, + "learning_rate": 0.0006, + "loss": 4.923543453216553, + "step": 5312 + }, + { + "epoch": 73.79650655021834, + "grad_norm": 0.028297102078795433, + "learning_rate": 0.0006, + "loss": 4.944367408752441, + "step": 5313 + }, + { + "epoch": 73.81048034934499, + "grad_norm": 0.027701659128069878, + "learning_rate": 0.0006, + "loss": 4.858168601989746, + "step": 5314 + }, + { + "epoch": 73.82445414847162, + "grad_norm": 0.02214639261364937, + "learning_rate": 0.0006, + "loss": 4.887052536010742, + "step": 5315 + }, + { + "epoch": 73.83842794759825, + "grad_norm": 0.01995093561708927, + "learning_rate": 0.0006, + "loss": 4.885797500610352, + "step": 5316 + }, + { + "epoch": 73.8524017467249, + "grad_norm": 0.020473985001444817, + "learning_rate": 0.0006, + "loss": 4.917325019836426, + "step": 5317 + }, + { + "epoch": 73.86637554585153, + "grad_norm": 0.0192868635058403, + "learning_rate": 0.0006, + "loss": 4.801746845245361, + "step": 5318 + }, + { + "epoch": 73.88034934497817, + "grad_norm": 0.017108267173171043, + "learning_rate": 0.0006, + "loss": 4.893620491027832, + "step": 5319 + }, + { + "epoch": 73.8943231441048, + "grad_norm": 0.01718071475625038, + "learning_rate": 0.0006, + "loss": 4.940695285797119, + "step": 5320 + }, + { + "epoch": 73.90829694323143, + "grad_norm": 0.017821557819843292, + "learning_rate": 0.0006, + "loss": 4.767613410949707, + "step": 5321 + }, + { + "epoch": 73.92227074235808, + "grad_norm": 0.015262950211763382, + "learning_rate": 0.0006, + "loss": 4.910315036773682, + "step": 5322 + }, + { + "epoch": 73.93624454148471, + "grad_norm": 0.01769006811082363, + "learning_rate": 0.0006, + "loss": 4.7743449211120605, + "step": 5323 + }, + { + "epoch": 73.95021834061136, + "grad_norm": 0.014460788108408451, + "learning_rate": 0.0006, + "loss": 4.743494987487793, + "step": 5324 + }, + { + "epoch": 73.96419213973799, + "grad_norm": 0.01403993833810091, + "learning_rate": 0.0006, + "loss": 4.666640281677246, + "step": 5325 + }, + { + "epoch": 73.97816593886463, + "grad_norm": 0.013997809961438179, + "learning_rate": 0.0006, + "loss": 4.698747158050537, + "step": 5326 + }, + { + "epoch": 73.99213973799127, + "grad_norm": 0.014244182966649532, + "learning_rate": 0.0006, + "loss": 4.745250701904297, + "step": 5327 + }, + { + "epoch": 74.0, + "grad_norm": 0.015144186094403267, + "learning_rate": 0.0006, + "loss": 4.667946815490723, + "step": 5328 + }, + { + "epoch": 74.0, + "eval_loss": 5.005440711975098, + "eval_runtime": 56.7836, + "eval_samples_per_second": 43.005, + "eval_steps_per_second": 1.356, + "step": 5328 + }, + { + "epoch": 74.01397379912663, + "grad_norm": 0.020384902134537697, + "learning_rate": 0.0006, + "loss": 4.626955509185791, + "step": 5329 + }, + { + "epoch": 74.02794759825328, + "grad_norm": 0.03624720871448517, + "learning_rate": 0.0006, + "loss": 4.713998317718506, + "step": 5330 + }, + { + "epoch": 74.04192139737991, + "grad_norm": 0.05701754242181778, + "learning_rate": 0.0006, + "loss": 4.717874526977539, + "step": 5331 + }, + { + "epoch": 74.05589519650655, + "grad_norm": 0.03942621126770973, + "learning_rate": 0.0006, + "loss": 4.7026190757751465, + "step": 5332 + }, + { + "epoch": 74.06986899563319, + "grad_norm": 0.017803272232413292, + "learning_rate": 0.0006, + "loss": 4.702548027038574, + "step": 5333 + }, + { + "epoch": 74.08384279475983, + "grad_norm": 0.017024677246809006, + "learning_rate": 0.0006, + "loss": 4.69752311706543, + "step": 5334 + }, + { + "epoch": 74.09781659388646, + "grad_norm": 0.016467854380607605, + "learning_rate": 0.0006, + "loss": 4.624876976013184, + "step": 5335 + }, + { + "epoch": 74.1117903930131, + "grad_norm": 0.016569215804338455, + "learning_rate": 0.0006, + "loss": 4.645040512084961, + "step": 5336 + }, + { + "epoch": 74.12576419213974, + "grad_norm": 0.012872067280113697, + "learning_rate": 0.0006, + "loss": 4.64582633972168, + "step": 5337 + }, + { + "epoch": 74.13973799126637, + "grad_norm": 0.015689387917518616, + "learning_rate": 0.0006, + "loss": 4.596040725708008, + "step": 5338 + }, + { + "epoch": 74.15371179039302, + "grad_norm": 0.013082082383334637, + "learning_rate": 0.0006, + "loss": 4.648033142089844, + "step": 5339 + }, + { + "epoch": 74.16768558951965, + "grad_norm": 0.013728172518312931, + "learning_rate": 0.0006, + "loss": 4.657873153686523, + "step": 5340 + }, + { + "epoch": 74.18165938864628, + "grad_norm": 0.013120594434440136, + "learning_rate": 0.0006, + "loss": 4.696074485778809, + "step": 5341 + }, + { + "epoch": 74.19563318777293, + "grad_norm": 0.01431245356798172, + "learning_rate": 0.0006, + "loss": 4.5679121017456055, + "step": 5342 + }, + { + "epoch": 74.20960698689956, + "grad_norm": 0.013994511216878891, + "learning_rate": 0.0006, + "loss": 4.5236406326293945, + "step": 5343 + }, + { + "epoch": 74.2235807860262, + "grad_norm": 0.011971707455813885, + "learning_rate": 0.0006, + "loss": 4.626547813415527, + "step": 5344 + }, + { + "epoch": 74.23755458515284, + "grad_norm": 0.012805287726223469, + "learning_rate": 0.0006, + "loss": 4.556386947631836, + "step": 5345 + }, + { + "epoch": 74.25152838427948, + "grad_norm": 0.011694584973156452, + "learning_rate": 0.0006, + "loss": 4.656987190246582, + "step": 5346 + }, + { + "epoch": 74.26550218340611, + "grad_norm": 0.013396821916103363, + "learning_rate": 0.0006, + "loss": 4.526467323303223, + "step": 5347 + }, + { + "epoch": 74.27947598253274, + "grad_norm": 0.014398031868040562, + "learning_rate": 0.0006, + "loss": 4.595100402832031, + "step": 5348 + }, + { + "epoch": 74.29344978165939, + "grad_norm": 0.02165800705552101, + "learning_rate": 0.0006, + "loss": 4.460549354553223, + "step": 5349 + }, + { + "epoch": 74.30742358078602, + "grad_norm": 0.03276512026786804, + "learning_rate": 0.0006, + "loss": 4.498189449310303, + "step": 5350 + }, + { + "epoch": 74.32139737991267, + "grad_norm": 0.04616224020719528, + "learning_rate": 0.0006, + "loss": 4.5268659591674805, + "step": 5351 + }, + { + "epoch": 74.3353711790393, + "grad_norm": 0.03863786533474922, + "learning_rate": 0.0006, + "loss": 4.554747104644775, + "step": 5352 + }, + { + "epoch": 74.34934497816595, + "grad_norm": 0.01749865710735321, + "learning_rate": 0.0006, + "loss": 4.6656036376953125, + "step": 5353 + }, + { + "epoch": 74.36331877729258, + "grad_norm": 0.021814396604895592, + "learning_rate": 0.0006, + "loss": 4.607456207275391, + "step": 5354 + }, + { + "epoch": 74.37729257641921, + "grad_norm": 0.017018496990203857, + "learning_rate": 0.0006, + "loss": 4.54205322265625, + "step": 5355 + }, + { + "epoch": 74.39126637554585, + "grad_norm": 0.018297340720891953, + "learning_rate": 0.0006, + "loss": 4.608870506286621, + "step": 5356 + }, + { + "epoch": 74.40524017467249, + "grad_norm": 0.01473482046276331, + "learning_rate": 0.0006, + "loss": 4.384891510009766, + "step": 5357 + }, + { + "epoch": 74.41921397379913, + "grad_norm": 0.01579415239393711, + "learning_rate": 0.0006, + "loss": 4.588859558105469, + "step": 5358 + }, + { + "epoch": 74.43318777292576, + "grad_norm": 0.013906026259064674, + "learning_rate": 0.0006, + "loss": 4.525175094604492, + "step": 5359 + }, + { + "epoch": 74.44716157205241, + "grad_norm": 0.014163361862301826, + "learning_rate": 0.0006, + "loss": 4.495802879333496, + "step": 5360 + }, + { + "epoch": 74.46113537117904, + "grad_norm": 0.014009335078299046, + "learning_rate": 0.0006, + "loss": 4.5960869789123535, + "step": 5361 + }, + { + "epoch": 74.47510917030567, + "grad_norm": 0.012765258550643921, + "learning_rate": 0.0006, + "loss": 4.515678405761719, + "step": 5362 + }, + { + "epoch": 74.48908296943232, + "grad_norm": 0.014071918092668056, + "learning_rate": 0.0006, + "loss": 4.394025802612305, + "step": 5363 + }, + { + "epoch": 74.50305676855895, + "grad_norm": 0.01244515273720026, + "learning_rate": 0.0006, + "loss": 4.576985836029053, + "step": 5364 + }, + { + "epoch": 74.5170305676856, + "grad_norm": 0.012975298799574375, + "learning_rate": 0.0006, + "loss": 4.540657997131348, + "step": 5365 + }, + { + "epoch": 74.53100436681223, + "grad_norm": 0.012380947358906269, + "learning_rate": 0.0006, + "loss": 4.55731725692749, + "step": 5366 + }, + { + "epoch": 74.54497816593886, + "grad_norm": 0.012577617540955544, + "learning_rate": 0.0006, + "loss": 4.500924110412598, + "step": 5367 + }, + { + "epoch": 74.5589519650655, + "grad_norm": 0.012089254334568977, + "learning_rate": 0.0006, + "loss": 4.561795234680176, + "step": 5368 + }, + { + "epoch": 74.57292576419214, + "grad_norm": 0.012024011462926865, + "learning_rate": 0.0006, + "loss": 4.430635452270508, + "step": 5369 + }, + { + "epoch": 74.58689956331878, + "grad_norm": 0.012361960485577583, + "learning_rate": 0.0006, + "loss": 4.4664106369018555, + "step": 5370 + }, + { + "epoch": 74.60087336244541, + "grad_norm": 0.012149893678724766, + "learning_rate": 0.0006, + "loss": 4.522271156311035, + "step": 5371 + }, + { + "epoch": 74.61484716157206, + "grad_norm": 0.01166840922087431, + "learning_rate": 0.0006, + "loss": 4.3870463371276855, + "step": 5372 + }, + { + "epoch": 74.62882096069869, + "grad_norm": 0.01160407718271017, + "learning_rate": 0.0006, + "loss": 4.504255771636963, + "step": 5373 + }, + { + "epoch": 74.64279475982532, + "grad_norm": 0.011259177699685097, + "learning_rate": 0.0006, + "loss": 4.507267951965332, + "step": 5374 + }, + { + "epoch": 74.65676855895197, + "grad_norm": 0.012285715900361538, + "learning_rate": 0.0006, + "loss": 4.472966194152832, + "step": 5375 + }, + { + "epoch": 74.6707423580786, + "grad_norm": 0.011214440688490868, + "learning_rate": 0.0006, + "loss": 4.5840253829956055, + "step": 5376 + }, + { + "epoch": 74.68471615720524, + "grad_norm": 0.012394879944622517, + "learning_rate": 0.0006, + "loss": 4.417830944061279, + "step": 5377 + }, + { + "epoch": 74.69868995633188, + "grad_norm": 0.012360049411654472, + "learning_rate": 0.0006, + "loss": 4.535454750061035, + "step": 5378 + }, + { + "epoch": 74.71266375545852, + "grad_norm": 0.011961652897298336, + "learning_rate": 0.0006, + "loss": 4.403409957885742, + "step": 5379 + }, + { + "epoch": 74.72663755458515, + "grad_norm": 0.012078274972736835, + "learning_rate": 0.0006, + "loss": 4.448366641998291, + "step": 5380 + }, + { + "epoch": 74.74061135371178, + "grad_norm": 0.013239333406090736, + "learning_rate": 0.0006, + "loss": 4.543450355529785, + "step": 5381 + }, + { + "epoch": 74.75458515283843, + "grad_norm": 0.01476441603153944, + "learning_rate": 0.0006, + "loss": 4.488677978515625, + "step": 5382 + }, + { + "epoch": 74.76855895196506, + "grad_norm": 0.013968009501695633, + "learning_rate": 0.0006, + "loss": 4.603639602661133, + "step": 5383 + }, + { + "epoch": 74.78253275109171, + "grad_norm": 0.012322458438575268, + "learning_rate": 0.0006, + "loss": 4.637874603271484, + "step": 5384 + }, + { + "epoch": 74.79650655021834, + "grad_norm": 0.013352593407034874, + "learning_rate": 0.0006, + "loss": 4.474833965301514, + "step": 5385 + }, + { + "epoch": 74.81048034934499, + "grad_norm": 0.014671728946268559, + "learning_rate": 0.0006, + "loss": 4.516847610473633, + "step": 5386 + }, + { + "epoch": 74.82445414847162, + "grad_norm": 0.017163321375846863, + "learning_rate": 0.0006, + "loss": 4.462325096130371, + "step": 5387 + }, + { + "epoch": 74.83842794759825, + "grad_norm": 0.02012511156499386, + "learning_rate": 0.0006, + "loss": 4.355982780456543, + "step": 5388 + }, + { + "epoch": 74.8524017467249, + "grad_norm": 0.01881697215139866, + "learning_rate": 0.0006, + "loss": 4.451120376586914, + "step": 5389 + }, + { + "epoch": 74.86637554585153, + "grad_norm": 0.017279407009482384, + "learning_rate": 0.0006, + "loss": 4.461202621459961, + "step": 5390 + }, + { + "epoch": 74.88034934497817, + "grad_norm": 0.015470407903194427, + "learning_rate": 0.0006, + "loss": 4.4968671798706055, + "step": 5391 + }, + { + "epoch": 74.8943231441048, + "grad_norm": 0.01411783043295145, + "learning_rate": 0.0006, + "loss": 4.558697700500488, + "step": 5392 + }, + { + "epoch": 74.90829694323143, + "grad_norm": 0.017178812995553017, + "learning_rate": 0.0006, + "loss": 4.254153251647949, + "step": 5393 + }, + { + "epoch": 74.92227074235808, + "grad_norm": 0.019193004816770554, + "learning_rate": 0.0006, + "loss": 4.394747257232666, + "step": 5394 + }, + { + "epoch": 74.93624454148471, + "grad_norm": 0.018942881375551224, + "learning_rate": 0.0006, + "loss": 4.459539413452148, + "step": 5395 + }, + { + "epoch": 74.95021834061136, + "grad_norm": 0.01641223020851612, + "learning_rate": 0.0006, + "loss": 4.485820770263672, + "step": 5396 + }, + { + "epoch": 74.96419213973799, + "grad_norm": 0.014485753141343594, + "learning_rate": 0.0006, + "loss": 4.5068583488464355, + "step": 5397 + }, + { + "epoch": 74.97816593886463, + "grad_norm": 0.018242299556732178, + "learning_rate": 0.0006, + "loss": 4.508835792541504, + "step": 5398 + }, + { + "epoch": 74.99213973799127, + "grad_norm": 0.016491873189806938, + "learning_rate": 0.0006, + "loss": 4.4760870933532715, + "step": 5399 + }, + { + "epoch": 75.0, + "grad_norm": 0.014543353579938412, + "learning_rate": 0.0006, + "loss": 4.540217399597168, + "step": 5400 + }, + { + "epoch": 75.0, + "eval_loss": 4.7617573738098145, + "eval_runtime": 57.321, + "eval_samples_per_second": 42.602, + "eval_steps_per_second": 1.343, + "step": 5400 + }, + { + "epoch": 75.01397379912663, + "grad_norm": 0.014417114667594433, + "learning_rate": 0.0006, + "loss": 4.332956790924072, + "step": 5401 + }, + { + "epoch": 75.02794759825328, + "grad_norm": 0.01607639715075493, + "learning_rate": 0.0006, + "loss": 4.352453708648682, + "step": 5402 + }, + { + "epoch": 75.04192139737991, + "grad_norm": 0.015438318252563477, + "learning_rate": 0.0006, + "loss": 4.387139320373535, + "step": 5403 + }, + { + "epoch": 75.05589519650655, + "grad_norm": 0.015539007261395454, + "learning_rate": 0.0006, + "loss": 4.453171730041504, + "step": 5404 + }, + { + "epoch": 75.06986899563319, + "grad_norm": 0.017361903563141823, + "learning_rate": 0.0006, + "loss": 4.512143135070801, + "step": 5405 + }, + { + "epoch": 75.08384279475983, + "grad_norm": 0.01815599389374256, + "learning_rate": 0.0006, + "loss": 4.508997440338135, + "step": 5406 + }, + { + "epoch": 75.09781659388646, + "grad_norm": 0.01620607078075409, + "learning_rate": 0.0006, + "loss": 4.322444915771484, + "step": 5407 + }, + { + "epoch": 75.1117903930131, + "grad_norm": 0.014773265458643436, + "learning_rate": 0.0006, + "loss": 4.359251976013184, + "step": 5408 + }, + { + "epoch": 75.12576419213974, + "grad_norm": 0.014099023304879665, + "learning_rate": 0.0006, + "loss": 4.313380241394043, + "step": 5409 + }, + { + "epoch": 75.13973799126637, + "grad_norm": 0.014387735165655613, + "learning_rate": 0.0006, + "loss": 4.435023307800293, + "step": 5410 + }, + { + "epoch": 75.15371179039302, + "grad_norm": 0.014886162243783474, + "learning_rate": 0.0006, + "loss": 4.335939407348633, + "step": 5411 + }, + { + "epoch": 75.16768558951965, + "grad_norm": 0.015766102820634842, + "learning_rate": 0.0006, + "loss": 4.408674240112305, + "step": 5412 + }, + { + "epoch": 75.18165938864628, + "grad_norm": 0.016383284702897072, + "learning_rate": 0.0006, + "loss": 4.385453224182129, + "step": 5413 + }, + { + "epoch": 75.19563318777293, + "grad_norm": 0.013535384088754654, + "learning_rate": 0.0006, + "loss": 4.292785167694092, + "step": 5414 + }, + { + "epoch": 75.20960698689956, + "grad_norm": 0.01354036945849657, + "learning_rate": 0.0006, + "loss": 4.306252479553223, + "step": 5415 + }, + { + "epoch": 75.2235807860262, + "grad_norm": 0.015818078070878983, + "learning_rate": 0.0006, + "loss": 4.425487995147705, + "step": 5416 + }, + { + "epoch": 75.23755458515284, + "grad_norm": 0.014363568276166916, + "learning_rate": 0.0006, + "loss": 4.395087242126465, + "step": 5417 + }, + { + "epoch": 75.25152838427948, + "grad_norm": 0.013895470649003983, + "learning_rate": 0.0006, + "loss": 4.364090442657471, + "step": 5418 + }, + { + "epoch": 75.26550218340611, + "grad_norm": 0.014178668148815632, + "learning_rate": 0.0006, + "loss": 4.44619083404541, + "step": 5419 + }, + { + "epoch": 75.27947598253274, + "grad_norm": 0.0152514623478055, + "learning_rate": 0.0006, + "loss": 4.348445892333984, + "step": 5420 + }, + { + "epoch": 75.29344978165939, + "grad_norm": 0.015397852286696434, + "learning_rate": 0.0006, + "loss": 4.476832389831543, + "step": 5421 + }, + { + "epoch": 75.30742358078602, + "grad_norm": 0.018009863793849945, + "learning_rate": 0.0006, + "loss": 4.431162357330322, + "step": 5422 + }, + { + "epoch": 75.32139737991267, + "grad_norm": 0.017038367688655853, + "learning_rate": 0.0006, + "loss": 4.474787712097168, + "step": 5423 + }, + { + "epoch": 75.3353711790393, + "grad_norm": 0.013763288035988808, + "learning_rate": 0.0006, + "loss": 4.470419883728027, + "step": 5424 + }, + { + "epoch": 75.34934497816595, + "grad_norm": 0.013972110114991665, + "learning_rate": 0.0006, + "loss": 4.373147964477539, + "step": 5425 + }, + { + "epoch": 75.36331877729258, + "grad_norm": 0.017197875306010246, + "learning_rate": 0.0006, + "loss": 4.494963645935059, + "step": 5426 + }, + { + "epoch": 75.37729257641921, + "grad_norm": 0.01687047816812992, + "learning_rate": 0.0006, + "loss": 4.435934066772461, + "step": 5427 + }, + { + "epoch": 75.39126637554585, + "grad_norm": 0.01578272506594658, + "learning_rate": 0.0006, + "loss": 4.37918758392334, + "step": 5428 + }, + { + "epoch": 75.40524017467249, + "grad_norm": 0.015956247225403786, + "learning_rate": 0.0006, + "loss": 4.455668926239014, + "step": 5429 + }, + { + "epoch": 75.41921397379913, + "grad_norm": 0.013495155610144138, + "learning_rate": 0.0006, + "loss": 4.4271039962768555, + "step": 5430 + }, + { + "epoch": 75.43318777292576, + "grad_norm": 0.012851621024310589, + "learning_rate": 0.0006, + "loss": 4.213107109069824, + "step": 5431 + }, + { + "epoch": 75.44716157205241, + "grad_norm": 0.013351158238947392, + "learning_rate": 0.0006, + "loss": 4.3070783615112305, + "step": 5432 + }, + { + "epoch": 75.46113537117904, + "grad_norm": 0.014765217900276184, + "learning_rate": 0.0006, + "loss": 4.385021209716797, + "step": 5433 + }, + { + "epoch": 75.47510917030567, + "grad_norm": 0.016091614961624146, + "learning_rate": 0.0006, + "loss": 4.312644004821777, + "step": 5434 + }, + { + "epoch": 75.48908296943232, + "grad_norm": 0.01612013578414917, + "learning_rate": 0.0006, + "loss": 4.4482011795043945, + "step": 5435 + }, + { + "epoch": 75.50305676855895, + "grad_norm": 0.013157603330910206, + "learning_rate": 0.0006, + "loss": 4.427262306213379, + "step": 5436 + }, + { + "epoch": 75.5170305676856, + "grad_norm": 0.016039809212088585, + "learning_rate": 0.0006, + "loss": 4.396635055541992, + "step": 5437 + }, + { + "epoch": 75.53100436681223, + "grad_norm": 0.01794283092021942, + "learning_rate": 0.0006, + "loss": 4.476870059967041, + "step": 5438 + }, + { + "epoch": 75.54497816593886, + "grad_norm": 0.017030486837029457, + "learning_rate": 0.0006, + "loss": 4.361156463623047, + "step": 5439 + }, + { + "epoch": 75.5589519650655, + "grad_norm": 0.017915161326527596, + "learning_rate": 0.0006, + "loss": 4.286147117614746, + "step": 5440 + }, + { + "epoch": 75.57292576419214, + "grad_norm": 0.017514433711767197, + "learning_rate": 0.0006, + "loss": 4.394091606140137, + "step": 5441 + }, + { + "epoch": 75.58689956331878, + "grad_norm": 0.015128589235246181, + "learning_rate": 0.0006, + "loss": 4.383126735687256, + "step": 5442 + }, + { + "epoch": 75.60087336244541, + "grad_norm": 0.0202616099268198, + "learning_rate": 0.0006, + "loss": 4.336188316345215, + "step": 5443 + }, + { + "epoch": 75.61484716157206, + "grad_norm": 0.019978471100330353, + "learning_rate": 0.0006, + "loss": 4.461453437805176, + "step": 5444 + }, + { + "epoch": 75.62882096069869, + "grad_norm": 0.01738741435110569, + "learning_rate": 0.0006, + "loss": 4.371197700500488, + "step": 5445 + }, + { + "epoch": 75.64279475982532, + "grad_norm": 0.016157809644937515, + "learning_rate": 0.0006, + "loss": 4.3312835693359375, + "step": 5446 + }, + { + "epoch": 75.65676855895197, + "grad_norm": 0.016394613310694695, + "learning_rate": 0.0006, + "loss": 4.380190849304199, + "step": 5447 + }, + { + "epoch": 75.6707423580786, + "grad_norm": 0.014727243222296238, + "learning_rate": 0.0006, + "loss": 4.308028221130371, + "step": 5448 + }, + { + "epoch": 75.68471615720524, + "grad_norm": 0.01695958711206913, + "learning_rate": 0.0006, + "loss": 4.399503707885742, + "step": 5449 + }, + { + "epoch": 75.69868995633188, + "grad_norm": 0.015720795840024948, + "learning_rate": 0.0006, + "loss": 4.342249870300293, + "step": 5450 + }, + { + "epoch": 75.71266375545852, + "grad_norm": 0.01574772410094738, + "learning_rate": 0.0006, + "loss": 4.331503868103027, + "step": 5451 + }, + { + "epoch": 75.72663755458515, + "grad_norm": 0.014812648296356201, + "learning_rate": 0.0006, + "loss": 4.332413673400879, + "step": 5452 + }, + { + "epoch": 75.74061135371178, + "grad_norm": 0.012850426137447357, + "learning_rate": 0.0006, + "loss": 4.2884063720703125, + "step": 5453 + }, + { + "epoch": 75.75458515283843, + "grad_norm": 0.012876970693469048, + "learning_rate": 0.0006, + "loss": 4.399855613708496, + "step": 5454 + }, + { + "epoch": 75.76855895196506, + "grad_norm": 0.012958863750100136, + "learning_rate": 0.0006, + "loss": 4.411185264587402, + "step": 5455 + }, + { + "epoch": 75.78253275109171, + "grad_norm": 0.012197135016322136, + "learning_rate": 0.0006, + "loss": 4.272603988647461, + "step": 5456 + }, + { + "epoch": 75.79650655021834, + "grad_norm": 0.013483566232025623, + "learning_rate": 0.0006, + "loss": 4.3564348220825195, + "step": 5457 + }, + { + "epoch": 75.81048034934499, + "grad_norm": 0.012365429662168026, + "learning_rate": 0.0006, + "loss": 4.354825973510742, + "step": 5458 + }, + { + "epoch": 75.82445414847162, + "grad_norm": 0.012238141149282455, + "learning_rate": 0.0006, + "loss": 4.440438747406006, + "step": 5459 + }, + { + "epoch": 75.83842794759825, + "grad_norm": 0.013380305841565132, + "learning_rate": 0.0006, + "loss": 4.2813215255737305, + "step": 5460 + }, + { + "epoch": 75.8524017467249, + "grad_norm": 0.01443779468536377, + "learning_rate": 0.0006, + "loss": 4.289078235626221, + "step": 5461 + }, + { + "epoch": 75.86637554585153, + "grad_norm": 0.012646574527025223, + "learning_rate": 0.0006, + "loss": 4.374177932739258, + "step": 5462 + }, + { + "epoch": 75.88034934497817, + "grad_norm": 0.012403388507664204, + "learning_rate": 0.0006, + "loss": 4.343016147613525, + "step": 5463 + }, + { + "epoch": 75.8943231441048, + "grad_norm": 0.01319252047687769, + "learning_rate": 0.0006, + "loss": 4.309957027435303, + "step": 5464 + }, + { + "epoch": 75.90829694323143, + "grad_norm": 0.012597428634762764, + "learning_rate": 0.0006, + "loss": 4.403654098510742, + "step": 5465 + }, + { + "epoch": 75.92227074235808, + "grad_norm": 0.012724226340651512, + "learning_rate": 0.0006, + "loss": 4.326869010925293, + "step": 5466 + }, + { + "epoch": 75.93624454148471, + "grad_norm": 0.012698772363364697, + "learning_rate": 0.0006, + "loss": 4.268430233001709, + "step": 5467 + }, + { + "epoch": 75.95021834061136, + "grad_norm": 0.01376490481197834, + "learning_rate": 0.0006, + "loss": 4.42446231842041, + "step": 5468 + }, + { + "epoch": 75.96419213973799, + "grad_norm": 0.013175144791603088, + "learning_rate": 0.0006, + "loss": 4.374675750732422, + "step": 5469 + }, + { + "epoch": 75.97816593886463, + "grad_norm": 0.012731322087347507, + "learning_rate": 0.0006, + "loss": 4.372966766357422, + "step": 5470 + }, + { + "epoch": 75.99213973799127, + "grad_norm": 0.01360505260527134, + "learning_rate": 0.0006, + "loss": 4.3095903396606445, + "step": 5471 + }, + { + "epoch": 76.0, + "grad_norm": 0.014011901803314686, + "learning_rate": 0.0006, + "loss": 4.26954460144043, + "step": 5472 + }, + { + "epoch": 76.0, + "eval_loss": 4.689640998840332, + "eval_runtime": 56.9095, + "eval_samples_per_second": 42.91, + "eval_steps_per_second": 1.353, + "step": 5472 + }, + { + "epoch": 76.01397379912663, + "grad_norm": 0.015351055189967155, + "learning_rate": 0.0006, + "loss": 4.348268985748291, + "step": 5473 + }, + { + "epoch": 76.02794759825328, + "grad_norm": 0.017164217308163643, + "learning_rate": 0.0006, + "loss": 4.244204998016357, + "step": 5474 + }, + { + "epoch": 76.04192139737991, + "grad_norm": 0.021624932065606117, + "learning_rate": 0.0006, + "loss": 4.311168670654297, + "step": 5475 + }, + { + "epoch": 76.05589519650655, + "grad_norm": 0.028712468221783638, + "learning_rate": 0.0006, + "loss": 4.322287082672119, + "step": 5476 + }, + { + "epoch": 76.06986899563319, + "grad_norm": 0.027915487065911293, + "learning_rate": 0.0006, + "loss": 4.381207466125488, + "step": 5477 + }, + { + "epoch": 76.08384279475983, + "grad_norm": 0.017955303192138672, + "learning_rate": 0.0006, + "loss": 4.2635016441345215, + "step": 5478 + }, + { + "epoch": 76.09781659388646, + "grad_norm": 0.019113514572381973, + "learning_rate": 0.0006, + "loss": 4.399471282958984, + "step": 5479 + }, + { + "epoch": 76.1117903930131, + "grad_norm": 0.017502542585134506, + "learning_rate": 0.0006, + "loss": 4.321204662322998, + "step": 5480 + }, + { + "epoch": 76.12576419213974, + "grad_norm": 0.01852525770664215, + "learning_rate": 0.0006, + "loss": 4.217884063720703, + "step": 5481 + }, + { + "epoch": 76.13973799126637, + "grad_norm": 0.019836828112602234, + "learning_rate": 0.0006, + "loss": 4.278532028198242, + "step": 5482 + }, + { + "epoch": 76.15371179039302, + "grad_norm": 0.02101975679397583, + "learning_rate": 0.0006, + "loss": 4.246720314025879, + "step": 5483 + }, + { + "epoch": 76.16768558951965, + "grad_norm": 0.021202990785241127, + "learning_rate": 0.0006, + "loss": 4.517517566680908, + "step": 5484 + }, + { + "epoch": 76.18165938864628, + "grad_norm": 0.021428769454360008, + "learning_rate": 0.0006, + "loss": 4.3373517990112305, + "step": 5485 + }, + { + "epoch": 76.19563318777293, + "grad_norm": 0.018263190984725952, + "learning_rate": 0.0006, + "loss": 4.402856826782227, + "step": 5486 + }, + { + "epoch": 76.20960698689956, + "grad_norm": 0.019025536254048347, + "learning_rate": 0.0006, + "loss": 4.314227104187012, + "step": 5487 + }, + { + "epoch": 76.2235807860262, + "grad_norm": 0.020918630063533783, + "learning_rate": 0.0006, + "loss": 4.454369068145752, + "step": 5488 + }, + { + "epoch": 76.23755458515284, + "grad_norm": 0.019307617098093033, + "learning_rate": 0.0006, + "loss": 4.247509002685547, + "step": 5489 + }, + { + "epoch": 76.25152838427948, + "grad_norm": 0.014956331811845303, + "learning_rate": 0.0006, + "loss": 4.365175247192383, + "step": 5490 + }, + { + "epoch": 76.26550218340611, + "grad_norm": 0.017213059589266777, + "learning_rate": 0.0006, + "loss": 4.384005069732666, + "step": 5491 + }, + { + "epoch": 76.27947598253274, + "grad_norm": 0.017575597390532494, + "learning_rate": 0.0006, + "loss": 4.41934871673584, + "step": 5492 + }, + { + "epoch": 76.29344978165939, + "grad_norm": 0.015879256650805473, + "learning_rate": 0.0006, + "loss": 4.26675271987915, + "step": 5493 + }, + { + "epoch": 76.30742358078602, + "grad_norm": 0.01564522460103035, + "learning_rate": 0.0006, + "loss": 4.277339935302734, + "step": 5494 + }, + { + "epoch": 76.32139737991267, + "grad_norm": 0.016414586454629898, + "learning_rate": 0.0006, + "loss": 4.4138593673706055, + "step": 5495 + }, + { + "epoch": 76.3353711790393, + "grad_norm": 0.014732337556779385, + "learning_rate": 0.0006, + "loss": 4.281006336212158, + "step": 5496 + }, + { + "epoch": 76.34934497816595, + "grad_norm": 0.015461455099284649, + "learning_rate": 0.0006, + "loss": 4.263422966003418, + "step": 5497 + }, + { + "epoch": 76.36331877729258, + "grad_norm": 0.014529814012348652, + "learning_rate": 0.0006, + "loss": 4.36165189743042, + "step": 5498 + }, + { + "epoch": 76.37729257641921, + "grad_norm": 0.01469483319669962, + "learning_rate": 0.0006, + "loss": 4.309642791748047, + "step": 5499 + }, + { + "epoch": 76.39126637554585, + "grad_norm": 0.016490574926137924, + "learning_rate": 0.0006, + "loss": 4.288150787353516, + "step": 5500 + }, + { + "epoch": 76.40524017467249, + "grad_norm": 0.014593565836548805, + "learning_rate": 0.0006, + "loss": 4.271614074707031, + "step": 5501 + }, + { + "epoch": 76.41921397379913, + "grad_norm": 0.012796535156667233, + "learning_rate": 0.0006, + "loss": 4.420042037963867, + "step": 5502 + }, + { + "epoch": 76.43318777292576, + "grad_norm": 0.016086025163531303, + "learning_rate": 0.0006, + "loss": 4.289943695068359, + "step": 5503 + }, + { + "epoch": 76.44716157205241, + "grad_norm": 0.015439470298588276, + "learning_rate": 0.0006, + "loss": 4.3023271560668945, + "step": 5504 + }, + { + "epoch": 76.46113537117904, + "grad_norm": 0.015066367574036121, + "learning_rate": 0.0006, + "loss": 4.197081565856934, + "step": 5505 + }, + { + "epoch": 76.47510917030567, + "grad_norm": 0.016653263941407204, + "learning_rate": 0.0006, + "loss": 4.333785057067871, + "step": 5506 + }, + { + "epoch": 76.48908296943232, + "grad_norm": 0.014791973866522312, + "learning_rate": 0.0006, + "loss": 4.323811054229736, + "step": 5507 + }, + { + "epoch": 76.50305676855895, + "grad_norm": 0.014694827608764172, + "learning_rate": 0.0006, + "loss": 4.195852279663086, + "step": 5508 + }, + { + "epoch": 76.5170305676856, + "grad_norm": 0.015120396390557289, + "learning_rate": 0.0006, + "loss": 4.42379903793335, + "step": 5509 + }, + { + "epoch": 76.53100436681223, + "grad_norm": 0.014584025368094444, + "learning_rate": 0.0006, + "loss": 4.2702531814575195, + "step": 5510 + }, + { + "epoch": 76.54497816593886, + "grad_norm": 0.013981613330543041, + "learning_rate": 0.0006, + "loss": 4.383029937744141, + "step": 5511 + }, + { + "epoch": 76.5589519650655, + "grad_norm": 0.012930642813444138, + "learning_rate": 0.0006, + "loss": 4.296136856079102, + "step": 5512 + }, + { + "epoch": 76.57292576419214, + "grad_norm": 0.013060989789664745, + "learning_rate": 0.0006, + "loss": 4.346380233764648, + "step": 5513 + }, + { + "epoch": 76.58689956331878, + "grad_norm": 0.013209798373281956, + "learning_rate": 0.0006, + "loss": 4.316067218780518, + "step": 5514 + }, + { + "epoch": 76.60087336244541, + "grad_norm": 0.012423216365277767, + "learning_rate": 0.0006, + "loss": 4.370274543762207, + "step": 5515 + }, + { + "epoch": 76.61484716157206, + "grad_norm": 0.013205411843955517, + "learning_rate": 0.0006, + "loss": 4.429969310760498, + "step": 5516 + }, + { + "epoch": 76.62882096069869, + "grad_norm": 0.013049827888607979, + "learning_rate": 0.0006, + "loss": 4.378549098968506, + "step": 5517 + }, + { + "epoch": 76.64279475982532, + "grad_norm": 0.012075880542397499, + "learning_rate": 0.0006, + "loss": 4.348293304443359, + "step": 5518 + }, + { + "epoch": 76.65676855895197, + "grad_norm": 0.014932847581803799, + "learning_rate": 0.0006, + "loss": 4.282538890838623, + "step": 5519 + }, + { + "epoch": 76.6707423580786, + "grad_norm": 0.015615998767316341, + "learning_rate": 0.0006, + "loss": 4.274641036987305, + "step": 5520 + }, + { + "epoch": 76.68471615720524, + "grad_norm": 0.01424513477832079, + "learning_rate": 0.0006, + "loss": 4.243166923522949, + "step": 5521 + }, + { + "epoch": 76.69868995633188, + "grad_norm": 0.012775546871125698, + "learning_rate": 0.0006, + "loss": 4.335611820220947, + "step": 5522 + }, + { + "epoch": 76.71266375545852, + "grad_norm": 0.014461303129792213, + "learning_rate": 0.0006, + "loss": 4.246976375579834, + "step": 5523 + }, + { + "epoch": 76.72663755458515, + "grad_norm": 0.0145436255261302, + "learning_rate": 0.0006, + "loss": 4.275895118713379, + "step": 5524 + }, + { + "epoch": 76.74061135371178, + "grad_norm": 0.014055085368454456, + "learning_rate": 0.0006, + "loss": 4.331108093261719, + "step": 5525 + }, + { + "epoch": 76.75458515283843, + "grad_norm": 0.011933894827961922, + "learning_rate": 0.0006, + "loss": 4.293721675872803, + "step": 5526 + }, + { + "epoch": 76.76855895196506, + "grad_norm": 0.01277392078191042, + "learning_rate": 0.0006, + "loss": 4.312374114990234, + "step": 5527 + }, + { + "epoch": 76.78253275109171, + "grad_norm": 0.014227936044335365, + "learning_rate": 0.0006, + "loss": 4.238729000091553, + "step": 5528 + }, + { + "epoch": 76.79650655021834, + "grad_norm": 0.01488024927675724, + "learning_rate": 0.0006, + "loss": 4.24587345123291, + "step": 5529 + }, + { + "epoch": 76.81048034934499, + "grad_norm": 0.015341250225901604, + "learning_rate": 0.0006, + "loss": 4.257378578186035, + "step": 5530 + }, + { + "epoch": 76.82445414847162, + "grad_norm": 0.015179437585175037, + "learning_rate": 0.0006, + "loss": 4.2812910079956055, + "step": 5531 + }, + { + "epoch": 76.83842794759825, + "grad_norm": 0.012814724817872047, + "learning_rate": 0.0006, + "loss": 4.2309465408325195, + "step": 5532 + }, + { + "epoch": 76.8524017467249, + "grad_norm": 0.013810716569423676, + "learning_rate": 0.0006, + "loss": 4.325292110443115, + "step": 5533 + }, + { + "epoch": 76.86637554585153, + "grad_norm": 0.017468232661485672, + "learning_rate": 0.0006, + "loss": 4.243788719177246, + "step": 5534 + }, + { + "epoch": 76.88034934497817, + "grad_norm": 0.01739857904613018, + "learning_rate": 0.0006, + "loss": 4.345390796661377, + "step": 5535 + }, + { + "epoch": 76.8943231441048, + "grad_norm": 0.015329061076045036, + "learning_rate": 0.0006, + "loss": 4.415142059326172, + "step": 5536 + }, + { + "epoch": 76.90829694323143, + "grad_norm": 0.01598426327109337, + "learning_rate": 0.0006, + "loss": 4.299558639526367, + "step": 5537 + }, + { + "epoch": 76.92227074235808, + "grad_norm": 0.01621454581618309, + "learning_rate": 0.0006, + "loss": 4.284354209899902, + "step": 5538 + }, + { + "epoch": 76.93624454148471, + "grad_norm": 0.015956968069076538, + "learning_rate": 0.0006, + "loss": 4.293837547302246, + "step": 5539 + }, + { + "epoch": 76.95021834061136, + "grad_norm": 0.01776493526995182, + "learning_rate": 0.0006, + "loss": 4.229729175567627, + "step": 5540 + }, + { + "epoch": 76.96419213973799, + "grad_norm": 0.015317477285861969, + "learning_rate": 0.0006, + "loss": 4.3328447341918945, + "step": 5541 + }, + { + "epoch": 76.97816593886463, + "grad_norm": 0.015140251256525517, + "learning_rate": 0.0006, + "loss": 4.310840606689453, + "step": 5542 + }, + { + "epoch": 76.99213973799127, + "grad_norm": 0.015326812863349915, + "learning_rate": 0.0006, + "loss": 4.230832099914551, + "step": 5543 + }, + { + "epoch": 77.0, + "grad_norm": 0.01607763208448887, + "learning_rate": 0.0006, + "loss": 4.2236328125, + "step": 5544 + }, + { + "epoch": 77.0, + "eval_loss": 4.662221431732178, + "eval_runtime": 56.5194, + "eval_samples_per_second": 43.206, + "eval_steps_per_second": 1.362, + "step": 5544 + }, + { + "epoch": 77.01397379912663, + "grad_norm": 0.014839510433375835, + "learning_rate": 0.0006, + "loss": 4.254111289978027, + "step": 5545 + }, + { + "epoch": 77.02794759825328, + "grad_norm": 0.01519741676747799, + "learning_rate": 0.0006, + "loss": 4.1607184410095215, + "step": 5546 + }, + { + "epoch": 77.04192139737991, + "grad_norm": 0.015472603030502796, + "learning_rate": 0.0006, + "loss": 4.323320388793945, + "step": 5547 + }, + { + "epoch": 77.05589519650655, + "grad_norm": 0.017937948927283287, + "learning_rate": 0.0006, + "loss": 4.308490753173828, + "step": 5548 + }, + { + "epoch": 77.06986899563319, + "grad_norm": 0.020890841260552406, + "learning_rate": 0.0006, + "loss": 4.259101867675781, + "step": 5549 + }, + { + "epoch": 77.08384279475983, + "grad_norm": 0.01892857998609543, + "learning_rate": 0.0006, + "loss": 4.386970520019531, + "step": 5550 + }, + { + "epoch": 77.09781659388646, + "grad_norm": 0.016583630815148354, + "learning_rate": 0.0006, + "loss": 4.233604907989502, + "step": 5551 + }, + { + "epoch": 77.1117903930131, + "grad_norm": 0.016063468530774117, + "learning_rate": 0.0006, + "loss": 4.339544296264648, + "step": 5552 + }, + { + "epoch": 77.12576419213974, + "grad_norm": 0.018326779827475548, + "learning_rate": 0.0006, + "loss": 4.367204666137695, + "step": 5553 + }, + { + "epoch": 77.13973799126637, + "grad_norm": 0.021590933203697205, + "learning_rate": 0.0006, + "loss": 4.352685928344727, + "step": 5554 + }, + { + "epoch": 77.15371179039302, + "grad_norm": 0.019643642008304596, + "learning_rate": 0.0006, + "loss": 4.258535385131836, + "step": 5555 + }, + { + "epoch": 77.16768558951965, + "grad_norm": 0.018476149067282677, + "learning_rate": 0.0006, + "loss": 4.232568740844727, + "step": 5556 + }, + { + "epoch": 77.18165938864628, + "grad_norm": 0.01584405079483986, + "learning_rate": 0.0006, + "loss": 4.258371353149414, + "step": 5557 + }, + { + "epoch": 77.19563318777293, + "grad_norm": 0.01575201191008091, + "learning_rate": 0.0006, + "loss": 4.183218002319336, + "step": 5558 + }, + { + "epoch": 77.20960698689956, + "grad_norm": 0.016871824860572815, + "learning_rate": 0.0006, + "loss": 4.194563865661621, + "step": 5559 + }, + { + "epoch": 77.2235807860262, + "grad_norm": 0.02059333398938179, + "learning_rate": 0.0006, + "loss": 4.195281982421875, + "step": 5560 + }, + { + "epoch": 77.23755458515284, + "grad_norm": 0.022640453651547432, + "learning_rate": 0.0006, + "loss": 4.279615879058838, + "step": 5561 + }, + { + "epoch": 77.25152838427948, + "grad_norm": 0.020335959270596504, + "learning_rate": 0.0006, + "loss": 4.25040864944458, + "step": 5562 + }, + { + "epoch": 77.26550218340611, + "grad_norm": 0.020010793581604958, + "learning_rate": 0.0006, + "loss": 4.291116714477539, + "step": 5563 + }, + { + "epoch": 77.27947598253274, + "grad_norm": 0.02109045907855034, + "learning_rate": 0.0006, + "loss": 4.364088535308838, + "step": 5564 + }, + { + "epoch": 77.29344978165939, + "grad_norm": 0.019519446417689323, + "learning_rate": 0.0006, + "loss": 4.329166412353516, + "step": 5565 + }, + { + "epoch": 77.30742358078602, + "grad_norm": 0.01689804159104824, + "learning_rate": 0.0006, + "loss": 4.244176864624023, + "step": 5566 + }, + { + "epoch": 77.32139737991267, + "grad_norm": 0.01893681474030018, + "learning_rate": 0.0006, + "loss": 4.239068984985352, + "step": 5567 + }, + { + "epoch": 77.3353711790393, + "grad_norm": 0.015998193994164467, + "learning_rate": 0.0006, + "loss": 4.173398017883301, + "step": 5568 + }, + { + "epoch": 77.34934497816595, + "grad_norm": 0.013892280869185925, + "learning_rate": 0.0006, + "loss": 4.287654876708984, + "step": 5569 + }, + { + "epoch": 77.36331877729258, + "grad_norm": 0.017060643061995506, + "learning_rate": 0.0006, + "loss": 4.398447036743164, + "step": 5570 + }, + { + "epoch": 77.37729257641921, + "grad_norm": 0.015168849378824234, + "learning_rate": 0.0006, + "loss": 4.15665864944458, + "step": 5571 + }, + { + "epoch": 77.39126637554585, + "grad_norm": 0.014886523596942425, + "learning_rate": 0.0006, + "loss": 4.196673393249512, + "step": 5572 + }, + { + "epoch": 77.40524017467249, + "grad_norm": 0.016941796988248825, + "learning_rate": 0.0006, + "loss": 4.314562797546387, + "step": 5573 + }, + { + "epoch": 77.41921397379913, + "grad_norm": 0.015353812836110592, + "learning_rate": 0.0006, + "loss": 4.409579277038574, + "step": 5574 + }, + { + "epoch": 77.43318777292576, + "grad_norm": 0.015192852355539799, + "learning_rate": 0.0006, + "loss": 4.392045974731445, + "step": 5575 + }, + { + "epoch": 77.44716157205241, + "grad_norm": 0.014232569374144077, + "learning_rate": 0.0006, + "loss": 4.21639347076416, + "step": 5576 + }, + { + "epoch": 77.46113537117904, + "grad_norm": 0.014453226700425148, + "learning_rate": 0.0006, + "loss": 4.2802276611328125, + "step": 5577 + }, + { + "epoch": 77.47510917030567, + "grad_norm": 0.012881236150860786, + "learning_rate": 0.0006, + "loss": 4.375980854034424, + "step": 5578 + }, + { + "epoch": 77.48908296943232, + "grad_norm": 0.01583976298570633, + "learning_rate": 0.0006, + "loss": 4.264520645141602, + "step": 5579 + }, + { + "epoch": 77.50305676855895, + "grad_norm": 0.014171584509313107, + "learning_rate": 0.0006, + "loss": 4.225279808044434, + "step": 5580 + }, + { + "epoch": 77.5170305676856, + "grad_norm": 0.014715159311890602, + "learning_rate": 0.0006, + "loss": 4.369834899902344, + "step": 5581 + }, + { + "epoch": 77.53100436681223, + "grad_norm": 0.016554079949855804, + "learning_rate": 0.0006, + "loss": 4.2976179122924805, + "step": 5582 + }, + { + "epoch": 77.54497816593886, + "grad_norm": 0.015095747075974941, + "learning_rate": 0.0006, + "loss": 4.294407844543457, + "step": 5583 + }, + { + "epoch": 77.5589519650655, + "grad_norm": 0.012726361863315105, + "learning_rate": 0.0006, + "loss": 4.221401214599609, + "step": 5584 + }, + { + "epoch": 77.57292576419214, + "grad_norm": 0.014999309554696083, + "learning_rate": 0.0006, + "loss": 4.243929386138916, + "step": 5585 + }, + { + "epoch": 77.58689956331878, + "grad_norm": 0.016526037827134132, + "learning_rate": 0.0006, + "loss": 4.374139785766602, + "step": 5586 + }, + { + "epoch": 77.60087336244541, + "grad_norm": 0.01648339442908764, + "learning_rate": 0.0006, + "loss": 4.392425060272217, + "step": 5587 + }, + { + "epoch": 77.61484716157206, + "grad_norm": 0.015269213356077671, + "learning_rate": 0.0006, + "loss": 4.33986759185791, + "step": 5588 + }, + { + "epoch": 77.62882096069869, + "grad_norm": 0.01690507121384144, + "learning_rate": 0.0006, + "loss": 4.307476043701172, + "step": 5589 + }, + { + "epoch": 77.64279475982532, + "grad_norm": 0.016438685357570648, + "learning_rate": 0.0006, + "loss": 4.2406206130981445, + "step": 5590 + }, + { + "epoch": 77.65676855895197, + "grad_norm": 0.0157600287348032, + "learning_rate": 0.0006, + "loss": 4.314070224761963, + "step": 5591 + }, + { + "epoch": 77.6707423580786, + "grad_norm": 0.015127947553992271, + "learning_rate": 0.0006, + "loss": 4.22475528717041, + "step": 5592 + }, + { + "epoch": 77.68471615720524, + "grad_norm": 0.016632677987217903, + "learning_rate": 0.0006, + "loss": 4.26954984664917, + "step": 5593 + }, + { + "epoch": 77.69868995633188, + "grad_norm": 0.013150460086762905, + "learning_rate": 0.0006, + "loss": 4.369131088256836, + "step": 5594 + }, + { + "epoch": 77.71266375545852, + "grad_norm": 0.014513005502521992, + "learning_rate": 0.0006, + "loss": 4.3248796463012695, + "step": 5595 + }, + { + "epoch": 77.72663755458515, + "grad_norm": 0.014090991578996181, + "learning_rate": 0.0006, + "loss": 4.268381118774414, + "step": 5596 + }, + { + "epoch": 77.74061135371178, + "grad_norm": 0.01633097417652607, + "learning_rate": 0.0006, + "loss": 4.238439083099365, + "step": 5597 + }, + { + "epoch": 77.75458515283843, + "grad_norm": 0.01710933819413185, + "learning_rate": 0.0006, + "loss": 4.399697780609131, + "step": 5598 + }, + { + "epoch": 77.76855895196506, + "grad_norm": 0.019251767545938492, + "learning_rate": 0.0006, + "loss": 4.290450096130371, + "step": 5599 + }, + { + "epoch": 77.78253275109171, + "grad_norm": 0.01849171705543995, + "learning_rate": 0.0006, + "loss": 4.267734527587891, + "step": 5600 + }, + { + "epoch": 77.79650655021834, + "grad_norm": 0.014173115603625774, + "learning_rate": 0.0006, + "loss": 4.288799285888672, + "step": 5601 + }, + { + "epoch": 77.81048034934499, + "grad_norm": 0.014572718180716038, + "learning_rate": 0.0006, + "loss": 4.3790435791015625, + "step": 5602 + }, + { + "epoch": 77.82445414847162, + "grad_norm": 0.016375336796045303, + "learning_rate": 0.0006, + "loss": 4.231725692749023, + "step": 5603 + }, + { + "epoch": 77.83842794759825, + "grad_norm": 0.017326701432466507, + "learning_rate": 0.0006, + "loss": 4.211495399475098, + "step": 5604 + }, + { + "epoch": 77.8524017467249, + "grad_norm": 0.016399390995502472, + "learning_rate": 0.0006, + "loss": 4.386699199676514, + "step": 5605 + }, + { + "epoch": 77.86637554585153, + "grad_norm": 0.018269989639520645, + "learning_rate": 0.0006, + "loss": 4.254856109619141, + "step": 5606 + }, + { + "epoch": 77.88034934497817, + "grad_norm": 0.018443435430526733, + "learning_rate": 0.0006, + "loss": 4.391961097717285, + "step": 5607 + }, + { + "epoch": 77.8943231441048, + "grad_norm": 0.01784636080265045, + "learning_rate": 0.0006, + "loss": 4.273881912231445, + "step": 5608 + }, + { + "epoch": 77.90829694323143, + "grad_norm": 0.01782994158565998, + "learning_rate": 0.0006, + "loss": 4.26833963394165, + "step": 5609 + }, + { + "epoch": 77.92227074235808, + "grad_norm": 0.016260741278529167, + "learning_rate": 0.0006, + "loss": 4.306807994842529, + "step": 5610 + }, + { + "epoch": 77.93624454148471, + "grad_norm": 0.015482903458178043, + "learning_rate": 0.0006, + "loss": 4.187826633453369, + "step": 5611 + }, + { + "epoch": 77.95021834061136, + "grad_norm": 0.015193257480859756, + "learning_rate": 0.0006, + "loss": 4.286060333251953, + "step": 5612 + }, + { + "epoch": 77.96419213973799, + "grad_norm": 0.015110115520656109, + "learning_rate": 0.0006, + "loss": 4.2828192710876465, + "step": 5613 + }, + { + "epoch": 77.97816593886463, + "grad_norm": 0.015973486006259918, + "learning_rate": 0.0006, + "loss": 4.317939758300781, + "step": 5614 + }, + { + "epoch": 77.99213973799127, + "grad_norm": 0.017363594844937325, + "learning_rate": 0.0006, + "loss": 4.287721633911133, + "step": 5615 + }, + { + "epoch": 78.0, + "grad_norm": 0.015277642756700516, + "learning_rate": 0.0006, + "loss": 4.356618881225586, + "step": 5616 + }, + { + "epoch": 78.0, + "eval_loss": 4.642678737640381, + "eval_runtime": 56.7335, + "eval_samples_per_second": 43.043, + "eval_steps_per_second": 1.357, + "step": 5616 + }, + { + "epoch": 78.01397379912663, + "grad_norm": 0.016794051975011826, + "learning_rate": 0.0006, + "loss": 4.279930591583252, + "step": 5617 + }, + { + "epoch": 78.02794759825328, + "grad_norm": 0.01632673107087612, + "learning_rate": 0.0006, + "loss": 4.235943794250488, + "step": 5618 + }, + { + "epoch": 78.04192139737991, + "grad_norm": 0.01437876746058464, + "learning_rate": 0.0006, + "loss": 4.2542243003845215, + "step": 5619 + }, + { + "epoch": 78.05589519650655, + "grad_norm": 0.014993119053542614, + "learning_rate": 0.0006, + "loss": 4.262046813964844, + "step": 5620 + }, + { + "epoch": 78.06986899563319, + "grad_norm": 0.015721119940280914, + "learning_rate": 0.0006, + "loss": 4.173436164855957, + "step": 5621 + }, + { + "epoch": 78.08384279475983, + "grad_norm": 0.01334085687994957, + "learning_rate": 0.0006, + "loss": 4.322175979614258, + "step": 5622 + }, + { + "epoch": 78.09781659388646, + "grad_norm": 0.015561388805508614, + "learning_rate": 0.0006, + "loss": 4.242125988006592, + "step": 5623 + }, + { + "epoch": 78.1117903930131, + "grad_norm": 0.014418602921068668, + "learning_rate": 0.0006, + "loss": 4.266186714172363, + "step": 5624 + }, + { + "epoch": 78.12576419213974, + "grad_norm": 0.013739909045398235, + "learning_rate": 0.0006, + "loss": 4.231633186340332, + "step": 5625 + }, + { + "epoch": 78.13973799126637, + "grad_norm": 0.01402067206799984, + "learning_rate": 0.0006, + "loss": 4.188406944274902, + "step": 5626 + }, + { + "epoch": 78.15371179039302, + "grad_norm": 0.013639481738209724, + "learning_rate": 0.0006, + "loss": 4.1586594581604, + "step": 5627 + }, + { + "epoch": 78.16768558951965, + "grad_norm": 0.012858159840106964, + "learning_rate": 0.0006, + "loss": 4.360448360443115, + "step": 5628 + }, + { + "epoch": 78.18165938864628, + "grad_norm": 0.013089739717543125, + "learning_rate": 0.0006, + "loss": 4.344371795654297, + "step": 5629 + }, + { + "epoch": 78.19563318777293, + "grad_norm": 0.01382314320653677, + "learning_rate": 0.0006, + "loss": 4.20378303527832, + "step": 5630 + }, + { + "epoch": 78.20960698689956, + "grad_norm": 0.014206619001924992, + "learning_rate": 0.0006, + "loss": 4.239891529083252, + "step": 5631 + }, + { + "epoch": 78.2235807860262, + "grad_norm": 0.01472168043255806, + "learning_rate": 0.0006, + "loss": 4.282382488250732, + "step": 5632 + }, + { + "epoch": 78.23755458515284, + "grad_norm": 0.014237025752663612, + "learning_rate": 0.0006, + "loss": 4.241545677185059, + "step": 5633 + }, + { + "epoch": 78.25152838427948, + "grad_norm": 0.013231472112238407, + "learning_rate": 0.0006, + "loss": 4.184896469116211, + "step": 5634 + }, + { + "epoch": 78.26550218340611, + "grad_norm": 0.013538471423089504, + "learning_rate": 0.0006, + "loss": 4.325181007385254, + "step": 5635 + }, + { + "epoch": 78.27947598253274, + "grad_norm": 0.013541961088776588, + "learning_rate": 0.0006, + "loss": 4.368722915649414, + "step": 5636 + }, + { + "epoch": 78.29344978165939, + "grad_norm": 0.013446723110973835, + "learning_rate": 0.0006, + "loss": 4.301422595977783, + "step": 5637 + }, + { + "epoch": 78.30742358078602, + "grad_norm": 0.014753809198737144, + "learning_rate": 0.0006, + "loss": 4.259127616882324, + "step": 5638 + }, + { + "epoch": 78.32139737991267, + "grad_norm": 0.013256008736789227, + "learning_rate": 0.0006, + "loss": 4.284032821655273, + "step": 5639 + }, + { + "epoch": 78.3353711790393, + "grad_norm": 0.014272629283368587, + "learning_rate": 0.0006, + "loss": 4.234989166259766, + "step": 5640 + }, + { + "epoch": 78.34934497816595, + "grad_norm": 0.014756826683878899, + "learning_rate": 0.0006, + "loss": 4.143499374389648, + "step": 5641 + }, + { + "epoch": 78.36331877729258, + "grad_norm": 0.016549205407500267, + "learning_rate": 0.0006, + "loss": 4.311034202575684, + "step": 5642 + }, + { + "epoch": 78.37729257641921, + "grad_norm": 0.017735283821821213, + "learning_rate": 0.0006, + "loss": 4.246767520904541, + "step": 5643 + }, + { + "epoch": 78.39126637554585, + "grad_norm": 0.017370613291859627, + "learning_rate": 0.0006, + "loss": 4.2460832595825195, + "step": 5644 + }, + { + "epoch": 78.40524017467249, + "grad_norm": 0.014642846770584583, + "learning_rate": 0.0006, + "loss": 4.303811550140381, + "step": 5645 + }, + { + "epoch": 78.41921397379913, + "grad_norm": 0.013445726595818996, + "learning_rate": 0.0006, + "loss": 4.25921630859375, + "step": 5646 + }, + { + "epoch": 78.43318777292576, + "grad_norm": 0.014254065230488777, + "learning_rate": 0.0006, + "loss": 4.129446506500244, + "step": 5647 + }, + { + "epoch": 78.44716157205241, + "grad_norm": 0.01883019506931305, + "learning_rate": 0.0006, + "loss": 4.191716194152832, + "step": 5648 + }, + { + "epoch": 78.46113537117904, + "grad_norm": 0.0204437468200922, + "learning_rate": 0.0006, + "loss": 4.333085536956787, + "step": 5649 + }, + { + "epoch": 78.47510917030567, + "grad_norm": 0.01652318798005581, + "learning_rate": 0.0006, + "loss": 4.135462760925293, + "step": 5650 + }, + { + "epoch": 78.48908296943232, + "grad_norm": 0.012753551825881004, + "learning_rate": 0.0006, + "loss": 4.195051193237305, + "step": 5651 + }, + { + "epoch": 78.50305676855895, + "grad_norm": 0.01499118935316801, + "learning_rate": 0.0006, + "loss": 4.157092094421387, + "step": 5652 + }, + { + "epoch": 78.5170305676856, + "grad_norm": 0.015383945778012276, + "learning_rate": 0.0006, + "loss": 4.253818035125732, + "step": 5653 + }, + { + "epoch": 78.53100436681223, + "grad_norm": 0.015721332281827927, + "learning_rate": 0.0006, + "loss": 4.228775978088379, + "step": 5654 + }, + { + "epoch": 78.54497816593886, + "grad_norm": 0.014595463871955872, + "learning_rate": 0.0006, + "loss": 4.135197639465332, + "step": 5655 + }, + { + "epoch": 78.5589519650655, + "grad_norm": 0.015059644356369972, + "learning_rate": 0.0006, + "loss": 4.255614280700684, + "step": 5656 + }, + { + "epoch": 78.57292576419214, + "grad_norm": 0.013833344914019108, + "learning_rate": 0.0006, + "loss": 4.301250457763672, + "step": 5657 + }, + { + "epoch": 78.58689956331878, + "grad_norm": 0.013665376231074333, + "learning_rate": 0.0006, + "loss": 4.1734232902526855, + "step": 5658 + }, + { + "epoch": 78.60087336244541, + "grad_norm": 0.01539086177945137, + "learning_rate": 0.0006, + "loss": 4.221459865570068, + "step": 5659 + }, + { + "epoch": 78.61484716157206, + "grad_norm": 0.013313917443156242, + "learning_rate": 0.0006, + "loss": 4.255619049072266, + "step": 5660 + }, + { + "epoch": 78.62882096069869, + "grad_norm": 0.01447741687297821, + "learning_rate": 0.0006, + "loss": 4.298219680786133, + "step": 5661 + }, + { + "epoch": 78.64279475982532, + "grad_norm": 0.012796915136277676, + "learning_rate": 0.0006, + "loss": 4.2542405128479, + "step": 5662 + }, + { + "epoch": 78.65676855895197, + "grad_norm": 0.014058141969144344, + "learning_rate": 0.0006, + "loss": 4.284477233886719, + "step": 5663 + }, + { + "epoch": 78.6707423580786, + "grad_norm": 0.012705069966614246, + "learning_rate": 0.0006, + "loss": 4.2826666831970215, + "step": 5664 + }, + { + "epoch": 78.68471615720524, + "grad_norm": 0.013245890848338604, + "learning_rate": 0.0006, + "loss": 4.371612548828125, + "step": 5665 + }, + { + "epoch": 78.69868995633188, + "grad_norm": 0.014192677102982998, + "learning_rate": 0.0006, + "loss": 4.267344951629639, + "step": 5666 + }, + { + "epoch": 78.71266375545852, + "grad_norm": 0.015346777625381947, + "learning_rate": 0.0006, + "loss": 4.222797393798828, + "step": 5667 + }, + { + "epoch": 78.72663755458515, + "grad_norm": 0.014068621210753918, + "learning_rate": 0.0006, + "loss": 4.172964572906494, + "step": 5668 + }, + { + "epoch": 78.74061135371178, + "grad_norm": 0.014364945702254772, + "learning_rate": 0.0006, + "loss": 4.259319305419922, + "step": 5669 + }, + { + "epoch": 78.75458515283843, + "grad_norm": 0.017337387427687645, + "learning_rate": 0.0006, + "loss": 4.29533576965332, + "step": 5670 + }, + { + "epoch": 78.76855895196506, + "grad_norm": 0.017416715621948242, + "learning_rate": 0.0006, + "loss": 4.340364456176758, + "step": 5671 + }, + { + "epoch": 78.78253275109171, + "grad_norm": 0.015925846993923187, + "learning_rate": 0.0006, + "loss": 4.177454948425293, + "step": 5672 + }, + { + "epoch": 78.79650655021834, + "grad_norm": 0.013130443170666695, + "learning_rate": 0.0006, + "loss": 4.277454853057861, + "step": 5673 + }, + { + "epoch": 78.81048034934499, + "grad_norm": 0.014628687873482704, + "learning_rate": 0.0006, + "loss": 4.273757457733154, + "step": 5674 + }, + { + "epoch": 78.82445414847162, + "grad_norm": 0.01601659320294857, + "learning_rate": 0.0006, + "loss": 4.281774520874023, + "step": 5675 + }, + { + "epoch": 78.83842794759825, + "grad_norm": 0.015443485230207443, + "learning_rate": 0.0006, + "loss": 4.199937343597412, + "step": 5676 + }, + { + "epoch": 78.8524017467249, + "grad_norm": 0.013934246264398098, + "learning_rate": 0.0006, + "loss": 4.331369876861572, + "step": 5677 + }, + { + "epoch": 78.86637554585153, + "grad_norm": 0.01572083681821823, + "learning_rate": 0.0006, + "loss": 4.2828474044799805, + "step": 5678 + }, + { + "epoch": 78.88034934497817, + "grad_norm": 0.013444120064377785, + "learning_rate": 0.0006, + "loss": 4.183859825134277, + "step": 5679 + }, + { + "epoch": 78.8943231441048, + "grad_norm": 0.01591254211962223, + "learning_rate": 0.0006, + "loss": 4.295256614685059, + "step": 5680 + }, + { + "epoch": 78.90829694323143, + "grad_norm": 0.017556684091687202, + "learning_rate": 0.0006, + "loss": 4.132272720336914, + "step": 5681 + }, + { + "epoch": 78.92227074235808, + "grad_norm": 0.018184322863817215, + "learning_rate": 0.0006, + "loss": 4.251728057861328, + "step": 5682 + }, + { + "epoch": 78.93624454148471, + "grad_norm": 0.016207285225391388, + "learning_rate": 0.0006, + "loss": 4.205544471740723, + "step": 5683 + }, + { + "epoch": 78.95021834061136, + "grad_norm": 0.014478376135230064, + "learning_rate": 0.0006, + "loss": 4.182462215423584, + "step": 5684 + }, + { + "epoch": 78.96419213973799, + "grad_norm": 0.014255235902965069, + "learning_rate": 0.0006, + "loss": 4.253366947174072, + "step": 5685 + }, + { + "epoch": 78.97816593886463, + "grad_norm": 0.01524396613240242, + "learning_rate": 0.0006, + "loss": 4.221786022186279, + "step": 5686 + }, + { + "epoch": 78.99213973799127, + "grad_norm": 0.015057693235576153, + "learning_rate": 0.0006, + "loss": 4.208454608917236, + "step": 5687 + }, + { + "epoch": 79.0, + "grad_norm": 0.015478022396564484, + "learning_rate": 0.0006, + "loss": 4.336383819580078, + "step": 5688 + }, + { + "epoch": 79.0, + "eval_loss": 4.641749382019043, + "eval_runtime": 56.8167, + "eval_samples_per_second": 42.98, + "eval_steps_per_second": 1.355, + "step": 5688 + }, + { + "epoch": 79.01397379912663, + "grad_norm": 0.014143316075205803, + "learning_rate": 0.0006, + "loss": 4.260974884033203, + "step": 5689 + }, + { + "epoch": 79.02794759825328, + "grad_norm": 0.013120883144438267, + "learning_rate": 0.0006, + "loss": 4.139886856079102, + "step": 5690 + }, + { + "epoch": 79.04192139737991, + "grad_norm": 0.015019661746919155, + "learning_rate": 0.0006, + "loss": 4.267004013061523, + "step": 5691 + }, + { + "epoch": 79.05589519650655, + "grad_norm": 0.014891284517943859, + "learning_rate": 0.0006, + "loss": 4.205019950866699, + "step": 5692 + }, + { + "epoch": 79.06986899563319, + "grad_norm": 0.015527728945016861, + "learning_rate": 0.0006, + "loss": 4.1513566970825195, + "step": 5693 + }, + { + "epoch": 79.08384279475983, + "grad_norm": 0.017527181655168533, + "learning_rate": 0.0006, + "loss": 4.148828983306885, + "step": 5694 + }, + { + "epoch": 79.09781659388646, + "grad_norm": 0.01632206328213215, + "learning_rate": 0.0006, + "loss": 4.195789813995361, + "step": 5695 + }, + { + "epoch": 79.1117903930131, + "grad_norm": 0.01590503193438053, + "learning_rate": 0.0006, + "loss": 4.199155330657959, + "step": 5696 + }, + { + "epoch": 79.12576419213974, + "grad_norm": 0.016489997506141663, + "learning_rate": 0.0006, + "loss": 4.155154705047607, + "step": 5697 + }, + { + "epoch": 79.13973799126637, + "grad_norm": 0.0161368977278471, + "learning_rate": 0.0006, + "loss": 4.164384841918945, + "step": 5698 + }, + { + "epoch": 79.15371179039302, + "grad_norm": 0.018574809655547142, + "learning_rate": 0.0006, + "loss": 4.1898193359375, + "step": 5699 + }, + { + "epoch": 79.16768558951965, + "grad_norm": 0.018987983465194702, + "learning_rate": 0.0006, + "loss": 4.219719409942627, + "step": 5700 + }, + { + "epoch": 79.18165938864628, + "grad_norm": 0.01668606884777546, + "learning_rate": 0.0006, + "loss": 4.152734756469727, + "step": 5701 + }, + { + "epoch": 79.19563318777293, + "grad_norm": 0.014046299271285534, + "learning_rate": 0.0006, + "loss": 4.152964115142822, + "step": 5702 + }, + { + "epoch": 79.20960698689956, + "grad_norm": 0.01628032699227333, + "learning_rate": 0.0006, + "loss": 4.299479961395264, + "step": 5703 + }, + { + "epoch": 79.2235807860262, + "grad_norm": 0.017998795956373215, + "learning_rate": 0.0006, + "loss": 4.231203079223633, + "step": 5704 + }, + { + "epoch": 79.23755458515284, + "grad_norm": 0.014799647964537144, + "learning_rate": 0.0006, + "loss": 4.3530192375183105, + "step": 5705 + }, + { + "epoch": 79.25152838427948, + "grad_norm": 0.014640678651630878, + "learning_rate": 0.0006, + "loss": 4.287322998046875, + "step": 5706 + }, + { + "epoch": 79.26550218340611, + "grad_norm": 0.0172616895288229, + "learning_rate": 0.0006, + "loss": 4.148009777069092, + "step": 5707 + }, + { + "epoch": 79.27947598253274, + "grad_norm": 0.01811790093779564, + "learning_rate": 0.0006, + "loss": 4.269433975219727, + "step": 5708 + }, + { + "epoch": 79.29344978165939, + "grad_norm": 0.01764000579714775, + "learning_rate": 0.0006, + "loss": 4.292638778686523, + "step": 5709 + }, + { + "epoch": 79.30742358078602, + "grad_norm": 0.018030935898423195, + "learning_rate": 0.0006, + "loss": 4.267889976501465, + "step": 5710 + }, + { + "epoch": 79.32139737991267, + "grad_norm": 0.017860891297459602, + "learning_rate": 0.0006, + "loss": 4.211110591888428, + "step": 5711 + }, + { + "epoch": 79.3353711790393, + "grad_norm": 0.017295073717832565, + "learning_rate": 0.0006, + "loss": 4.224645614624023, + "step": 5712 + }, + { + "epoch": 79.34934497816595, + "grad_norm": 0.015891101211309433, + "learning_rate": 0.0006, + "loss": 4.224797248840332, + "step": 5713 + }, + { + "epoch": 79.36331877729258, + "grad_norm": 0.014680417254567146, + "learning_rate": 0.0006, + "loss": 4.182242393493652, + "step": 5714 + }, + { + "epoch": 79.37729257641921, + "grad_norm": 0.016621140763163567, + "learning_rate": 0.0006, + "loss": 4.186334609985352, + "step": 5715 + }, + { + "epoch": 79.39126637554585, + "grad_norm": 0.016207249835133553, + "learning_rate": 0.0006, + "loss": 4.277074813842773, + "step": 5716 + }, + { + "epoch": 79.40524017467249, + "grad_norm": 0.015375478193163872, + "learning_rate": 0.0006, + "loss": 4.154317855834961, + "step": 5717 + }, + { + "epoch": 79.41921397379913, + "grad_norm": 0.013453228399157524, + "learning_rate": 0.0006, + "loss": 4.265270233154297, + "step": 5718 + }, + { + "epoch": 79.43318777292576, + "grad_norm": 0.014312694780528545, + "learning_rate": 0.0006, + "loss": 4.232220649719238, + "step": 5719 + }, + { + "epoch": 79.44716157205241, + "grad_norm": 0.014759554527699947, + "learning_rate": 0.0006, + "loss": 4.272275924682617, + "step": 5720 + }, + { + "epoch": 79.46113537117904, + "grad_norm": 0.015885505825281143, + "learning_rate": 0.0006, + "loss": 4.2229719161987305, + "step": 5721 + }, + { + "epoch": 79.47510917030567, + "grad_norm": 0.016516581177711487, + "learning_rate": 0.0006, + "loss": 4.200624465942383, + "step": 5722 + }, + { + "epoch": 79.48908296943232, + "grad_norm": 0.018254555761814117, + "learning_rate": 0.0006, + "loss": 4.19648551940918, + "step": 5723 + }, + { + "epoch": 79.50305676855895, + "grad_norm": 0.01467090006917715, + "learning_rate": 0.0006, + "loss": 4.135190486907959, + "step": 5724 + }, + { + "epoch": 79.5170305676856, + "grad_norm": 0.013474797829985619, + "learning_rate": 0.0006, + "loss": 4.311361312866211, + "step": 5725 + }, + { + "epoch": 79.53100436681223, + "grad_norm": 0.01742764748632908, + "learning_rate": 0.0006, + "loss": 4.1845903396606445, + "step": 5726 + }, + { + "epoch": 79.54497816593886, + "grad_norm": 0.018786882981657982, + "learning_rate": 0.0006, + "loss": 4.163942337036133, + "step": 5727 + }, + { + "epoch": 79.5589519650655, + "grad_norm": 0.017831219360232353, + "learning_rate": 0.0006, + "loss": 4.236685276031494, + "step": 5728 + }, + { + "epoch": 79.57292576419214, + "grad_norm": 0.016334986314177513, + "learning_rate": 0.0006, + "loss": 4.255853652954102, + "step": 5729 + }, + { + "epoch": 79.58689956331878, + "grad_norm": 0.0138487434014678, + "learning_rate": 0.0006, + "loss": 4.227874279022217, + "step": 5730 + }, + { + "epoch": 79.60087336244541, + "grad_norm": 0.01531760673969984, + "learning_rate": 0.0006, + "loss": 4.221432685852051, + "step": 5731 + }, + { + "epoch": 79.61484716157206, + "grad_norm": 0.01603361964225769, + "learning_rate": 0.0006, + "loss": 4.278326511383057, + "step": 5732 + }, + { + "epoch": 79.62882096069869, + "grad_norm": 0.013149895705282688, + "learning_rate": 0.0006, + "loss": 4.2474517822265625, + "step": 5733 + }, + { + "epoch": 79.64279475982532, + "grad_norm": 0.013428786769509315, + "learning_rate": 0.0006, + "loss": 4.1701178550720215, + "step": 5734 + }, + { + "epoch": 79.65676855895197, + "grad_norm": 0.014305585995316505, + "learning_rate": 0.0006, + "loss": 4.250778675079346, + "step": 5735 + }, + { + "epoch": 79.6707423580786, + "grad_norm": 0.014544566161930561, + "learning_rate": 0.0006, + "loss": 4.243236541748047, + "step": 5736 + }, + { + "epoch": 79.68471615720524, + "grad_norm": 0.013524414040148258, + "learning_rate": 0.0006, + "loss": 4.255457878112793, + "step": 5737 + }, + { + "epoch": 79.69868995633188, + "grad_norm": 0.014750641770660877, + "learning_rate": 0.0006, + "loss": 4.202816486358643, + "step": 5738 + }, + { + "epoch": 79.71266375545852, + "grad_norm": 0.015090773813426495, + "learning_rate": 0.0006, + "loss": 4.3053178787231445, + "step": 5739 + }, + { + "epoch": 79.72663755458515, + "grad_norm": 0.01606169529259205, + "learning_rate": 0.0006, + "loss": 4.224017143249512, + "step": 5740 + }, + { + "epoch": 79.74061135371178, + "grad_norm": 0.016303520649671555, + "learning_rate": 0.0006, + "loss": 4.190760612487793, + "step": 5741 + }, + { + "epoch": 79.75458515283843, + "grad_norm": 0.01465672068297863, + "learning_rate": 0.0006, + "loss": 4.24506950378418, + "step": 5742 + }, + { + "epoch": 79.76855895196506, + "grad_norm": 0.012616301886737347, + "learning_rate": 0.0006, + "loss": 4.174039840698242, + "step": 5743 + }, + { + "epoch": 79.78253275109171, + "grad_norm": 0.01358284242451191, + "learning_rate": 0.0006, + "loss": 4.236380577087402, + "step": 5744 + }, + { + "epoch": 79.79650655021834, + "grad_norm": 0.01557177770882845, + "learning_rate": 0.0006, + "loss": 4.102962493896484, + "step": 5745 + }, + { + "epoch": 79.81048034934499, + "grad_norm": 0.015584317035973072, + "learning_rate": 0.0006, + "loss": 4.242603302001953, + "step": 5746 + }, + { + "epoch": 79.82445414847162, + "grad_norm": 0.012597334571182728, + "learning_rate": 0.0006, + "loss": 4.202791690826416, + "step": 5747 + }, + { + "epoch": 79.83842794759825, + "grad_norm": 0.013722898438572884, + "learning_rate": 0.0006, + "loss": 4.204328536987305, + "step": 5748 + }, + { + "epoch": 79.8524017467249, + "grad_norm": 0.014669225551187992, + "learning_rate": 0.0006, + "loss": 4.164331912994385, + "step": 5749 + }, + { + "epoch": 79.86637554585153, + "grad_norm": 0.015012740157544613, + "learning_rate": 0.0006, + "loss": 4.260565280914307, + "step": 5750 + }, + { + "epoch": 79.88034934497817, + "grad_norm": 0.016026340425014496, + "learning_rate": 0.0006, + "loss": 4.299723148345947, + "step": 5751 + }, + { + "epoch": 79.8943231441048, + "grad_norm": 0.01626049168407917, + "learning_rate": 0.0006, + "loss": 4.207252502441406, + "step": 5752 + }, + { + "epoch": 79.90829694323143, + "grad_norm": 0.015415354631841183, + "learning_rate": 0.0006, + "loss": 4.151934623718262, + "step": 5753 + }, + { + "epoch": 79.92227074235808, + "grad_norm": 0.014158242382109165, + "learning_rate": 0.0006, + "loss": 4.3461995124816895, + "step": 5754 + }, + { + "epoch": 79.93624454148471, + "grad_norm": 0.01585628278553486, + "learning_rate": 0.0006, + "loss": 4.334761619567871, + "step": 5755 + }, + { + "epoch": 79.95021834061136, + "grad_norm": 0.013710759580135345, + "learning_rate": 0.0006, + "loss": 4.180401802062988, + "step": 5756 + }, + { + "epoch": 79.96419213973799, + "grad_norm": 0.014159204438328743, + "learning_rate": 0.0006, + "loss": 4.322331428527832, + "step": 5757 + }, + { + "epoch": 79.97816593886463, + "grad_norm": 0.015153195708990097, + "learning_rate": 0.0006, + "loss": 4.287478923797607, + "step": 5758 + }, + { + "epoch": 79.99213973799127, + "grad_norm": 0.014100227504968643, + "learning_rate": 0.0006, + "loss": 4.231807708740234, + "step": 5759 + }, + { + "epoch": 80.0, + "grad_norm": 0.015248682349920273, + "learning_rate": 0.0006, + "loss": 4.261106014251709, + "step": 5760 + }, + { + "epoch": 80.0, + "eval_loss": 4.610264301300049, + "eval_runtime": 56.514, + "eval_samples_per_second": 43.211, + "eval_steps_per_second": 1.362, + "step": 5760 + }, + { + "epoch": 80.01397379912663, + "grad_norm": 0.014730271883308887, + "learning_rate": 0.0006, + "loss": 4.108556747436523, + "step": 5761 + }, + { + "epoch": 80.02794759825328, + "grad_norm": 0.014575716108083725, + "learning_rate": 0.0006, + "loss": 4.169658660888672, + "step": 5762 + }, + { + "epoch": 80.04192139737991, + "grad_norm": 0.014706958085298538, + "learning_rate": 0.0006, + "loss": 4.252662658691406, + "step": 5763 + }, + { + "epoch": 80.05589519650655, + "grad_norm": 0.01561224739998579, + "learning_rate": 0.0006, + "loss": 4.193892955780029, + "step": 5764 + }, + { + "epoch": 80.06986899563319, + "grad_norm": 0.014702324755489826, + "learning_rate": 0.0006, + "loss": 4.1889238357543945, + "step": 5765 + }, + { + "epoch": 80.08384279475983, + "grad_norm": 0.015565741807222366, + "learning_rate": 0.0006, + "loss": 4.243549823760986, + "step": 5766 + }, + { + "epoch": 80.09781659388646, + "grad_norm": 0.01647181063890457, + "learning_rate": 0.0006, + "loss": 4.275446891784668, + "step": 5767 + }, + { + "epoch": 80.1117903930131, + "grad_norm": 0.018177716061472893, + "learning_rate": 0.0006, + "loss": 4.12040376663208, + "step": 5768 + }, + { + "epoch": 80.12576419213974, + "grad_norm": 0.01745191030204296, + "learning_rate": 0.0006, + "loss": 4.140371322631836, + "step": 5769 + }, + { + "epoch": 80.13973799126637, + "grad_norm": 0.01568993367254734, + "learning_rate": 0.0006, + "loss": 4.258280277252197, + "step": 5770 + }, + { + "epoch": 80.15371179039302, + "grad_norm": 0.01636597141623497, + "learning_rate": 0.0006, + "loss": 4.179052352905273, + "step": 5771 + }, + { + "epoch": 80.16768558951965, + "grad_norm": 0.016839100047945976, + "learning_rate": 0.0006, + "loss": 4.127636432647705, + "step": 5772 + }, + { + "epoch": 80.18165938864628, + "grad_norm": 0.015471016988158226, + "learning_rate": 0.0006, + "loss": 4.181325912475586, + "step": 5773 + }, + { + "epoch": 80.19563318777293, + "grad_norm": 0.015836089849472046, + "learning_rate": 0.0006, + "loss": 4.142168998718262, + "step": 5774 + }, + { + "epoch": 80.20960698689956, + "grad_norm": 0.014782954007387161, + "learning_rate": 0.0006, + "loss": 4.308799743652344, + "step": 5775 + }, + { + "epoch": 80.2235807860262, + "grad_norm": 0.014040893875062466, + "learning_rate": 0.0006, + "loss": 4.216936111450195, + "step": 5776 + }, + { + "epoch": 80.23755458515284, + "grad_norm": 0.01420280709862709, + "learning_rate": 0.0006, + "loss": 4.235230445861816, + "step": 5777 + }, + { + "epoch": 80.25152838427948, + "grad_norm": 0.013341007754206657, + "learning_rate": 0.0006, + "loss": 4.369109153747559, + "step": 5778 + }, + { + "epoch": 80.26550218340611, + "grad_norm": 0.014208327978849411, + "learning_rate": 0.0006, + "loss": 4.219112396240234, + "step": 5779 + }, + { + "epoch": 80.27947598253274, + "grad_norm": 0.013825052417814732, + "learning_rate": 0.0006, + "loss": 4.1983489990234375, + "step": 5780 + }, + { + "epoch": 80.29344978165939, + "grad_norm": 0.012916515581309795, + "learning_rate": 0.0006, + "loss": 4.204582691192627, + "step": 5781 + }, + { + "epoch": 80.30742358078602, + "grad_norm": 0.013071243651211262, + "learning_rate": 0.0006, + "loss": 4.17643928527832, + "step": 5782 + }, + { + "epoch": 80.32139737991267, + "grad_norm": 0.015507174655795097, + "learning_rate": 0.0006, + "loss": 4.172976016998291, + "step": 5783 + }, + { + "epoch": 80.3353711790393, + "grad_norm": 0.014967243187129498, + "learning_rate": 0.0006, + "loss": 4.141168594360352, + "step": 5784 + }, + { + "epoch": 80.34934497816595, + "grad_norm": 0.015629924833774567, + "learning_rate": 0.0006, + "loss": 4.253526210784912, + "step": 5785 + }, + { + "epoch": 80.36331877729258, + "grad_norm": 0.01607954502105713, + "learning_rate": 0.0006, + "loss": 4.249395370483398, + "step": 5786 + }, + { + "epoch": 80.37729257641921, + "grad_norm": 0.01500706560909748, + "learning_rate": 0.0006, + "loss": 4.279012680053711, + "step": 5787 + }, + { + "epoch": 80.39126637554585, + "grad_norm": 0.014343900606036186, + "learning_rate": 0.0006, + "loss": 4.268509864807129, + "step": 5788 + }, + { + "epoch": 80.40524017467249, + "grad_norm": 0.014682197012007236, + "learning_rate": 0.0006, + "loss": 4.206334114074707, + "step": 5789 + }, + { + "epoch": 80.41921397379913, + "grad_norm": 0.014989163726568222, + "learning_rate": 0.0006, + "loss": 4.143601417541504, + "step": 5790 + }, + { + "epoch": 80.43318777292576, + "grad_norm": 0.015001608058810234, + "learning_rate": 0.0006, + "loss": 4.164111137390137, + "step": 5791 + }, + { + "epoch": 80.44716157205241, + "grad_norm": 0.015222722664475441, + "learning_rate": 0.0006, + "loss": 4.2369914054870605, + "step": 5792 + }, + { + "epoch": 80.46113537117904, + "grad_norm": 0.014126183465123177, + "learning_rate": 0.0006, + "loss": 4.2209930419921875, + "step": 5793 + }, + { + "epoch": 80.47510917030567, + "grad_norm": 0.013065959326922894, + "learning_rate": 0.0006, + "loss": 4.231184959411621, + "step": 5794 + }, + { + "epoch": 80.48908296943232, + "grad_norm": 0.013833962380886078, + "learning_rate": 0.0006, + "loss": 4.226434707641602, + "step": 5795 + }, + { + "epoch": 80.50305676855895, + "grad_norm": 0.01387125626206398, + "learning_rate": 0.0006, + "loss": 4.241328716278076, + "step": 5796 + }, + { + "epoch": 80.5170305676856, + "grad_norm": 0.014145063236355782, + "learning_rate": 0.0006, + "loss": 4.2268805503845215, + "step": 5797 + }, + { + "epoch": 80.53100436681223, + "grad_norm": 0.012990240007638931, + "learning_rate": 0.0006, + "loss": 4.1718339920043945, + "step": 5798 + }, + { + "epoch": 80.54497816593886, + "grad_norm": 0.014185411855578423, + "learning_rate": 0.0006, + "loss": 4.232397556304932, + "step": 5799 + }, + { + "epoch": 80.5589519650655, + "grad_norm": 0.014260523952543736, + "learning_rate": 0.0006, + "loss": 4.282734394073486, + "step": 5800 + }, + { + "epoch": 80.57292576419214, + "grad_norm": 0.012542120181024075, + "learning_rate": 0.0006, + "loss": 4.134551525115967, + "step": 5801 + }, + { + "epoch": 80.58689956331878, + "grad_norm": 0.013078657910227776, + "learning_rate": 0.0006, + "loss": 4.229815483093262, + "step": 5802 + }, + { + "epoch": 80.60087336244541, + "grad_norm": 0.012821056880056858, + "learning_rate": 0.0006, + "loss": 4.208278179168701, + "step": 5803 + }, + { + "epoch": 80.61484716157206, + "grad_norm": 0.012964180670678616, + "learning_rate": 0.0006, + "loss": 4.188477039337158, + "step": 5804 + }, + { + "epoch": 80.62882096069869, + "grad_norm": 0.01631746254861355, + "learning_rate": 0.0006, + "loss": 4.3214497566223145, + "step": 5805 + }, + { + "epoch": 80.64279475982532, + "grad_norm": 0.01726640947163105, + "learning_rate": 0.0006, + "loss": 4.279359817504883, + "step": 5806 + }, + { + "epoch": 80.65676855895197, + "grad_norm": 0.018778258934617043, + "learning_rate": 0.0006, + "loss": 4.294047832489014, + "step": 5807 + }, + { + "epoch": 80.6707423580786, + "grad_norm": 0.01796831749379635, + "learning_rate": 0.0006, + "loss": 4.314162254333496, + "step": 5808 + }, + { + "epoch": 80.68471615720524, + "grad_norm": 0.016495538875460625, + "learning_rate": 0.0006, + "loss": 4.273871421813965, + "step": 5809 + }, + { + "epoch": 80.69868995633188, + "grad_norm": 0.013590306043624878, + "learning_rate": 0.0006, + "loss": 4.14961051940918, + "step": 5810 + }, + { + "epoch": 80.71266375545852, + "grad_norm": 0.014670426957309246, + "learning_rate": 0.0006, + "loss": 4.211853981018066, + "step": 5811 + }, + { + "epoch": 80.72663755458515, + "grad_norm": 0.01641341857612133, + "learning_rate": 0.0006, + "loss": 4.211948871612549, + "step": 5812 + }, + { + "epoch": 80.74061135371178, + "grad_norm": 0.016365470364689827, + "learning_rate": 0.0006, + "loss": 4.200774192810059, + "step": 5813 + }, + { + "epoch": 80.75458515283843, + "grad_norm": 0.01613970287144184, + "learning_rate": 0.0006, + "loss": 4.288322448730469, + "step": 5814 + }, + { + "epoch": 80.76855895196506, + "grad_norm": 0.016742777079343796, + "learning_rate": 0.0006, + "loss": 4.23801326751709, + "step": 5815 + }, + { + "epoch": 80.78253275109171, + "grad_norm": 0.015351069159805775, + "learning_rate": 0.0006, + "loss": 4.263832092285156, + "step": 5816 + }, + { + "epoch": 80.79650655021834, + "grad_norm": 0.013435336761176586, + "learning_rate": 0.0006, + "loss": 4.160857200622559, + "step": 5817 + }, + { + "epoch": 80.81048034934499, + "grad_norm": 0.01474886853247881, + "learning_rate": 0.0006, + "loss": 4.216025352478027, + "step": 5818 + }, + { + "epoch": 80.82445414847162, + "grad_norm": 0.01586935855448246, + "learning_rate": 0.0006, + "loss": 4.256402969360352, + "step": 5819 + }, + { + "epoch": 80.83842794759825, + "grad_norm": 0.014596572145819664, + "learning_rate": 0.0006, + "loss": 4.142584800720215, + "step": 5820 + }, + { + "epoch": 80.8524017467249, + "grad_norm": 0.01501593366265297, + "learning_rate": 0.0006, + "loss": 4.250462055206299, + "step": 5821 + }, + { + "epoch": 80.86637554585153, + "grad_norm": 0.014773511327803135, + "learning_rate": 0.0006, + "loss": 4.253305912017822, + "step": 5822 + }, + { + "epoch": 80.88034934497817, + "grad_norm": 0.013246326707303524, + "learning_rate": 0.0006, + "loss": 4.170881271362305, + "step": 5823 + }, + { + "epoch": 80.8943231441048, + "grad_norm": 0.013013756833970547, + "learning_rate": 0.0006, + "loss": 4.264721870422363, + "step": 5824 + }, + { + "epoch": 80.90829694323143, + "grad_norm": 0.013155903667211533, + "learning_rate": 0.0006, + "loss": 4.274667739868164, + "step": 5825 + }, + { + "epoch": 80.92227074235808, + "grad_norm": 0.01448952779173851, + "learning_rate": 0.0006, + "loss": 4.24777889251709, + "step": 5826 + }, + { + "epoch": 80.93624454148471, + "grad_norm": 0.01415231078863144, + "learning_rate": 0.0006, + "loss": 4.22014856338501, + "step": 5827 + }, + { + "epoch": 80.95021834061136, + "grad_norm": 0.014796273782849312, + "learning_rate": 0.0006, + "loss": 4.22036600112915, + "step": 5828 + }, + { + "epoch": 80.96419213973799, + "grad_norm": 0.013005326502025127, + "learning_rate": 0.0006, + "loss": 4.276123046875, + "step": 5829 + }, + { + "epoch": 80.97816593886463, + "grad_norm": 0.012851177714765072, + "learning_rate": 0.0006, + "loss": 4.248575687408447, + "step": 5830 + }, + { + "epoch": 80.99213973799127, + "grad_norm": 0.013740002177655697, + "learning_rate": 0.0006, + "loss": 4.051140308380127, + "step": 5831 + }, + { + "epoch": 81.0, + "grad_norm": 0.015296104364097118, + "learning_rate": 0.0006, + "loss": 4.251155853271484, + "step": 5832 + }, + { + "epoch": 81.0, + "eval_loss": 4.594127655029297, + "eval_runtime": 56.0959, + "eval_samples_per_second": 43.533, + "eval_steps_per_second": 1.373, + "step": 5832 + }, + { + "epoch": 81.01397379912663, + "grad_norm": 0.015804558992385864, + "learning_rate": 0.0006, + "loss": 4.152894020080566, + "step": 5833 + }, + { + "epoch": 81.02794759825328, + "grad_norm": 0.017000854015350342, + "learning_rate": 0.0006, + "loss": 4.266067981719971, + "step": 5834 + }, + { + "epoch": 81.04192139737991, + "grad_norm": 0.016351008787751198, + "learning_rate": 0.0006, + "loss": 4.192703723907471, + "step": 5835 + }, + { + "epoch": 81.05589519650655, + "grad_norm": 0.014703895896673203, + "learning_rate": 0.0006, + "loss": 4.23554801940918, + "step": 5836 + }, + { + "epoch": 81.06986899563319, + "grad_norm": 0.01619596965610981, + "learning_rate": 0.0006, + "loss": 4.233000755310059, + "step": 5837 + }, + { + "epoch": 81.08384279475983, + "grad_norm": 0.01815493032336235, + "learning_rate": 0.0006, + "loss": 4.200490951538086, + "step": 5838 + }, + { + "epoch": 81.09781659388646, + "grad_norm": 0.018305106088519096, + "learning_rate": 0.0006, + "loss": 4.154242515563965, + "step": 5839 + }, + { + "epoch": 81.1117903930131, + "grad_norm": 0.0200856514275074, + "learning_rate": 0.0006, + "loss": 4.25820779800415, + "step": 5840 + }, + { + "epoch": 81.12576419213974, + "grad_norm": 0.02101440541446209, + "learning_rate": 0.0006, + "loss": 4.168209075927734, + "step": 5841 + }, + { + "epoch": 81.13973799126637, + "grad_norm": 0.022829681634902954, + "learning_rate": 0.0006, + "loss": 4.25560998916626, + "step": 5842 + }, + { + "epoch": 81.15371179039302, + "grad_norm": 0.0195082426071167, + "learning_rate": 0.0006, + "loss": 4.2035627365112305, + "step": 5843 + }, + { + "epoch": 81.16768558951965, + "grad_norm": 0.01716248318552971, + "learning_rate": 0.0006, + "loss": 4.11322021484375, + "step": 5844 + }, + { + "epoch": 81.18165938864628, + "grad_norm": 0.02034664712846279, + "learning_rate": 0.0006, + "loss": 4.179389953613281, + "step": 5845 + }, + { + "epoch": 81.19563318777293, + "grad_norm": 0.023578710854053497, + "learning_rate": 0.0006, + "loss": 4.265695571899414, + "step": 5846 + }, + { + "epoch": 81.20960698689956, + "grad_norm": 0.019244182854890823, + "learning_rate": 0.0006, + "loss": 4.292043209075928, + "step": 5847 + }, + { + "epoch": 81.2235807860262, + "grad_norm": 0.01577291637659073, + "learning_rate": 0.0006, + "loss": 4.250776290893555, + "step": 5848 + }, + { + "epoch": 81.23755458515284, + "grad_norm": 0.01617090217769146, + "learning_rate": 0.0006, + "loss": 4.193098068237305, + "step": 5849 + }, + { + "epoch": 81.25152838427948, + "grad_norm": 0.016090184450149536, + "learning_rate": 0.0006, + "loss": 4.112357139587402, + "step": 5850 + }, + { + "epoch": 81.26550218340611, + "grad_norm": 0.01441365759819746, + "learning_rate": 0.0006, + "loss": 4.114147186279297, + "step": 5851 + }, + { + "epoch": 81.27947598253274, + "grad_norm": 0.014122162945568562, + "learning_rate": 0.0006, + "loss": 4.276175022125244, + "step": 5852 + }, + { + "epoch": 81.29344978165939, + "grad_norm": 0.014859487302601337, + "learning_rate": 0.0006, + "loss": 4.094699859619141, + "step": 5853 + }, + { + "epoch": 81.30742358078602, + "grad_norm": 0.013624267652630806, + "learning_rate": 0.0006, + "loss": 4.2648773193359375, + "step": 5854 + }, + { + "epoch": 81.32139737991267, + "grad_norm": 0.01417913381010294, + "learning_rate": 0.0006, + "loss": 4.228567123413086, + "step": 5855 + }, + { + "epoch": 81.3353711790393, + "grad_norm": 0.014560805633664131, + "learning_rate": 0.0006, + "loss": 4.148930549621582, + "step": 5856 + }, + { + "epoch": 81.34934497816595, + "grad_norm": 0.013662154786288738, + "learning_rate": 0.0006, + "loss": 4.192107677459717, + "step": 5857 + }, + { + "epoch": 81.36331877729258, + "grad_norm": 0.013434980995953083, + "learning_rate": 0.0006, + "loss": 4.162052154541016, + "step": 5858 + }, + { + "epoch": 81.37729257641921, + "grad_norm": 0.01300742756575346, + "learning_rate": 0.0006, + "loss": 4.262385368347168, + "step": 5859 + }, + { + "epoch": 81.39126637554585, + "grad_norm": 0.012544351629912853, + "learning_rate": 0.0006, + "loss": 4.215371131896973, + "step": 5860 + }, + { + "epoch": 81.40524017467249, + "grad_norm": 0.013663645833730698, + "learning_rate": 0.0006, + "loss": 4.247312545776367, + "step": 5861 + }, + { + "epoch": 81.41921397379913, + "grad_norm": 0.013127296231687069, + "learning_rate": 0.0006, + "loss": 4.087685585021973, + "step": 5862 + }, + { + "epoch": 81.43318777292576, + "grad_norm": 0.014081938192248344, + "learning_rate": 0.0006, + "loss": 4.272648811340332, + "step": 5863 + }, + { + "epoch": 81.44716157205241, + "grad_norm": 0.015470641665160656, + "learning_rate": 0.0006, + "loss": 4.278578758239746, + "step": 5864 + }, + { + "epoch": 81.46113537117904, + "grad_norm": 0.016524532809853554, + "learning_rate": 0.0006, + "loss": 4.118268013000488, + "step": 5865 + }, + { + "epoch": 81.47510917030567, + "grad_norm": 0.016242722049355507, + "learning_rate": 0.0006, + "loss": 4.182405471801758, + "step": 5866 + }, + { + "epoch": 81.48908296943232, + "grad_norm": 0.01557017583400011, + "learning_rate": 0.0006, + "loss": 4.229862213134766, + "step": 5867 + }, + { + "epoch": 81.50305676855895, + "grad_norm": 0.01562870666384697, + "learning_rate": 0.0006, + "loss": 4.27754020690918, + "step": 5868 + }, + { + "epoch": 81.5170305676856, + "grad_norm": 0.018606074154376984, + "learning_rate": 0.0006, + "loss": 4.148681640625, + "step": 5869 + }, + { + "epoch": 81.53100436681223, + "grad_norm": 0.017740730196237564, + "learning_rate": 0.0006, + "loss": 4.321932792663574, + "step": 5870 + }, + { + "epoch": 81.54497816593886, + "grad_norm": 0.014517677947878838, + "learning_rate": 0.0006, + "loss": 4.22862434387207, + "step": 5871 + }, + { + "epoch": 81.5589519650655, + "grad_norm": 0.01389714702963829, + "learning_rate": 0.0006, + "loss": 4.198000907897949, + "step": 5872 + }, + { + "epoch": 81.57292576419214, + "grad_norm": 0.015348682180047035, + "learning_rate": 0.0006, + "loss": 4.293485164642334, + "step": 5873 + }, + { + "epoch": 81.58689956331878, + "grad_norm": 0.015965791419148445, + "learning_rate": 0.0006, + "loss": 4.23481559753418, + "step": 5874 + }, + { + "epoch": 81.60087336244541, + "grad_norm": 0.015662726014852524, + "learning_rate": 0.0006, + "loss": 4.158148765563965, + "step": 5875 + }, + { + "epoch": 81.61484716157206, + "grad_norm": 0.014402080327272415, + "learning_rate": 0.0006, + "loss": 4.132723808288574, + "step": 5876 + }, + { + "epoch": 81.62882096069869, + "grad_norm": 0.016846125945448875, + "learning_rate": 0.0006, + "loss": 4.1654462814331055, + "step": 5877 + }, + { + "epoch": 81.64279475982532, + "grad_norm": 0.0196097232401371, + "learning_rate": 0.0006, + "loss": 4.228450298309326, + "step": 5878 + }, + { + "epoch": 81.65676855895197, + "grad_norm": 0.016005294397473335, + "learning_rate": 0.0006, + "loss": 4.230222225189209, + "step": 5879 + }, + { + "epoch": 81.6707423580786, + "grad_norm": 0.014685059897601604, + "learning_rate": 0.0006, + "loss": 4.1367387771606445, + "step": 5880 + }, + { + "epoch": 81.68471615720524, + "grad_norm": 0.01647144928574562, + "learning_rate": 0.0006, + "loss": 4.096675395965576, + "step": 5881 + }, + { + "epoch": 81.69868995633188, + "grad_norm": 0.015978610143065453, + "learning_rate": 0.0006, + "loss": 4.232292175292969, + "step": 5882 + }, + { + "epoch": 81.71266375545852, + "grad_norm": 0.01510288193821907, + "learning_rate": 0.0006, + "loss": 4.274591445922852, + "step": 5883 + }, + { + "epoch": 81.72663755458515, + "grad_norm": 0.014942476525902748, + "learning_rate": 0.0006, + "loss": 4.161482810974121, + "step": 5884 + }, + { + "epoch": 81.74061135371178, + "grad_norm": 0.014595299027860165, + "learning_rate": 0.0006, + "loss": 4.204291343688965, + "step": 5885 + }, + { + "epoch": 81.75458515283843, + "grad_norm": 0.014952429570257664, + "learning_rate": 0.0006, + "loss": 4.242499351501465, + "step": 5886 + }, + { + "epoch": 81.76855895196506, + "grad_norm": 0.014761554077267647, + "learning_rate": 0.0006, + "loss": 4.20393180847168, + "step": 5887 + }, + { + "epoch": 81.78253275109171, + "grad_norm": 0.014749647118151188, + "learning_rate": 0.0006, + "loss": 4.227565765380859, + "step": 5888 + }, + { + "epoch": 81.79650655021834, + "grad_norm": 0.014384493231773376, + "learning_rate": 0.0006, + "loss": 4.282228946685791, + "step": 5889 + }, + { + "epoch": 81.81048034934499, + "grad_norm": 0.01260330155491829, + "learning_rate": 0.0006, + "loss": 4.2882866859436035, + "step": 5890 + }, + { + "epoch": 81.82445414847162, + "grad_norm": 0.014055311679840088, + "learning_rate": 0.0006, + "loss": 4.22351598739624, + "step": 5891 + }, + { + "epoch": 81.83842794759825, + "grad_norm": 0.013674065470695496, + "learning_rate": 0.0006, + "loss": 4.240716934204102, + "step": 5892 + }, + { + "epoch": 81.8524017467249, + "grad_norm": 0.01225196197628975, + "learning_rate": 0.0006, + "loss": 4.2958598136901855, + "step": 5893 + }, + { + "epoch": 81.86637554585153, + "grad_norm": 0.013538680970668793, + "learning_rate": 0.0006, + "loss": 4.24342679977417, + "step": 5894 + }, + { + "epoch": 81.88034934497817, + "grad_norm": 0.013151394203305244, + "learning_rate": 0.0006, + "loss": 4.081999778747559, + "step": 5895 + }, + { + "epoch": 81.8943231441048, + "grad_norm": 0.014260678552091122, + "learning_rate": 0.0006, + "loss": 4.156423568725586, + "step": 5896 + }, + { + "epoch": 81.90829694323143, + "grad_norm": 0.01329776830971241, + "learning_rate": 0.0006, + "loss": 4.1713995933532715, + "step": 5897 + }, + { + "epoch": 81.92227074235808, + "grad_norm": 0.014623276889324188, + "learning_rate": 0.0006, + "loss": 4.244891166687012, + "step": 5898 + }, + { + "epoch": 81.93624454148471, + "grad_norm": 0.015760324895381927, + "learning_rate": 0.0006, + "loss": 4.076528549194336, + "step": 5899 + }, + { + "epoch": 81.95021834061136, + "grad_norm": 0.014331766404211521, + "learning_rate": 0.0006, + "loss": 4.283037185668945, + "step": 5900 + }, + { + "epoch": 81.96419213973799, + "grad_norm": 0.012383842840790749, + "learning_rate": 0.0006, + "loss": 4.117308616638184, + "step": 5901 + }, + { + "epoch": 81.97816593886463, + "grad_norm": 0.01250660140067339, + "learning_rate": 0.0006, + "loss": 4.241250991821289, + "step": 5902 + }, + { + "epoch": 81.99213973799127, + "grad_norm": 0.012276900932192802, + "learning_rate": 0.0006, + "loss": 4.141091823577881, + "step": 5903 + }, + { + "epoch": 82.0, + "grad_norm": 0.01588737778365612, + "learning_rate": 0.0006, + "loss": 4.065532684326172, + "step": 5904 + }, + { + "epoch": 82.0, + "eval_loss": 4.612304210662842, + "eval_runtime": 57.5411, + "eval_samples_per_second": 42.439, + "eval_steps_per_second": 1.338, + "step": 5904 + }, + { + "epoch": 82.01397379912663, + "grad_norm": 0.017433544620871544, + "learning_rate": 0.0006, + "loss": 4.202497482299805, + "step": 5905 + }, + { + "epoch": 82.02794759825328, + "grad_norm": 0.017260415479540825, + "learning_rate": 0.0006, + "loss": 4.099949359893799, + "step": 5906 + }, + { + "epoch": 82.04192139737991, + "grad_norm": 0.01577908545732498, + "learning_rate": 0.0006, + "loss": 4.17806339263916, + "step": 5907 + }, + { + "epoch": 82.05589519650655, + "grad_norm": 0.016093818470835686, + "learning_rate": 0.0006, + "loss": 4.219491004943848, + "step": 5908 + }, + { + "epoch": 82.06986899563319, + "grad_norm": 0.015426302328705788, + "learning_rate": 0.0006, + "loss": 4.207982063293457, + "step": 5909 + }, + { + "epoch": 82.08384279475983, + "grad_norm": 0.01605442352592945, + "learning_rate": 0.0006, + "loss": 4.265942573547363, + "step": 5910 + }, + { + "epoch": 82.09781659388646, + "grad_norm": 0.01787225715816021, + "learning_rate": 0.0006, + "loss": 4.151535511016846, + "step": 5911 + }, + { + "epoch": 82.1117903930131, + "grad_norm": 0.01750164106488228, + "learning_rate": 0.0006, + "loss": 4.133655071258545, + "step": 5912 + }, + { + "epoch": 82.12576419213974, + "grad_norm": 0.01602957211434841, + "learning_rate": 0.0006, + "loss": 4.176972389221191, + "step": 5913 + }, + { + "epoch": 82.13973799126637, + "grad_norm": 0.01553812064230442, + "learning_rate": 0.0006, + "loss": 4.125692367553711, + "step": 5914 + }, + { + "epoch": 82.15371179039302, + "grad_norm": 0.013978508301079273, + "learning_rate": 0.0006, + "loss": 4.207271575927734, + "step": 5915 + }, + { + "epoch": 82.16768558951965, + "grad_norm": 0.013561118394136429, + "learning_rate": 0.0006, + "loss": 4.238930702209473, + "step": 5916 + }, + { + "epoch": 82.18165938864628, + "grad_norm": 0.013285507448017597, + "learning_rate": 0.0006, + "loss": 4.129464626312256, + "step": 5917 + }, + { + "epoch": 82.19563318777293, + "grad_norm": 0.013903839513659477, + "learning_rate": 0.0006, + "loss": 4.214109420776367, + "step": 5918 + }, + { + "epoch": 82.20960698689956, + "grad_norm": 0.014426837675273418, + "learning_rate": 0.0006, + "loss": 4.143196105957031, + "step": 5919 + }, + { + "epoch": 82.2235807860262, + "grad_norm": 0.015347165986895561, + "learning_rate": 0.0006, + "loss": 4.235439300537109, + "step": 5920 + }, + { + "epoch": 82.23755458515284, + "grad_norm": 0.014266039244830608, + "learning_rate": 0.0006, + "loss": 3.991617202758789, + "step": 5921 + }, + { + "epoch": 82.25152838427948, + "grad_norm": 0.013101593591272831, + "learning_rate": 0.0006, + "loss": 4.266017913818359, + "step": 5922 + }, + { + "epoch": 82.26550218340611, + "grad_norm": 0.013741881586611271, + "learning_rate": 0.0006, + "loss": 4.224879741668701, + "step": 5923 + }, + { + "epoch": 82.27947598253274, + "grad_norm": 0.014206945896148682, + "learning_rate": 0.0006, + "loss": 4.102272987365723, + "step": 5924 + }, + { + "epoch": 82.29344978165939, + "grad_norm": 0.013798723928630352, + "learning_rate": 0.0006, + "loss": 4.30418586730957, + "step": 5925 + }, + { + "epoch": 82.30742358078602, + "grad_norm": 0.015470122918486595, + "learning_rate": 0.0006, + "loss": 4.2513322830200195, + "step": 5926 + }, + { + "epoch": 82.32139737991267, + "grad_norm": 0.01655442826449871, + "learning_rate": 0.0006, + "loss": 4.223002910614014, + "step": 5927 + }, + { + "epoch": 82.3353711790393, + "grad_norm": 0.017170004546642303, + "learning_rate": 0.0006, + "loss": 4.253904819488525, + "step": 5928 + }, + { + "epoch": 82.34934497816595, + "grad_norm": 0.017595812678337097, + "learning_rate": 0.0006, + "loss": 4.184927940368652, + "step": 5929 + }, + { + "epoch": 82.36331877729258, + "grad_norm": 0.015073256567120552, + "learning_rate": 0.0006, + "loss": 4.202239036560059, + "step": 5930 + }, + { + "epoch": 82.37729257641921, + "grad_norm": 0.014310941100120544, + "learning_rate": 0.0006, + "loss": 4.124934673309326, + "step": 5931 + }, + { + "epoch": 82.39126637554585, + "grad_norm": 0.013671821914613247, + "learning_rate": 0.0006, + "loss": 4.138919353485107, + "step": 5932 + }, + { + "epoch": 82.40524017467249, + "grad_norm": 0.014103314839303493, + "learning_rate": 0.0006, + "loss": 4.211289405822754, + "step": 5933 + }, + { + "epoch": 82.41921397379913, + "grad_norm": 0.014272668398916721, + "learning_rate": 0.0006, + "loss": 4.06290340423584, + "step": 5934 + }, + { + "epoch": 82.43318777292576, + "grad_norm": 0.01387016475200653, + "learning_rate": 0.0006, + "loss": 4.101334571838379, + "step": 5935 + }, + { + "epoch": 82.44716157205241, + "grad_norm": 0.013318683952093124, + "learning_rate": 0.0006, + "loss": 4.206409931182861, + "step": 5936 + }, + { + "epoch": 82.46113537117904, + "grad_norm": 0.013781268149614334, + "learning_rate": 0.0006, + "loss": 4.189474105834961, + "step": 5937 + }, + { + "epoch": 82.47510917030567, + "grad_norm": 0.013195851817727089, + "learning_rate": 0.0006, + "loss": 4.276607036590576, + "step": 5938 + }, + { + "epoch": 82.48908296943232, + "grad_norm": 0.014358163811266422, + "learning_rate": 0.0006, + "loss": 4.203262805938721, + "step": 5939 + }, + { + "epoch": 82.50305676855895, + "grad_norm": 0.01403636671602726, + "learning_rate": 0.0006, + "loss": 4.079258441925049, + "step": 5940 + }, + { + "epoch": 82.5170305676856, + "grad_norm": 0.013272535055875778, + "learning_rate": 0.0006, + "loss": 4.19197940826416, + "step": 5941 + }, + { + "epoch": 82.53100436681223, + "grad_norm": 0.013787264935672283, + "learning_rate": 0.0006, + "loss": 4.12995719909668, + "step": 5942 + }, + { + "epoch": 82.54497816593886, + "grad_norm": 0.015020057559013367, + "learning_rate": 0.0006, + "loss": 4.148890495300293, + "step": 5943 + }, + { + "epoch": 82.5589519650655, + "grad_norm": 0.014199157245457172, + "learning_rate": 0.0006, + "loss": 4.201516151428223, + "step": 5944 + }, + { + "epoch": 82.57292576419214, + "grad_norm": 0.013827123679220676, + "learning_rate": 0.0006, + "loss": 4.240894317626953, + "step": 5945 + }, + { + "epoch": 82.58689956331878, + "grad_norm": 0.014767560176551342, + "learning_rate": 0.0006, + "loss": 4.173320770263672, + "step": 5946 + }, + { + "epoch": 82.60087336244541, + "grad_norm": 0.01612078957259655, + "learning_rate": 0.0006, + "loss": 4.266448020935059, + "step": 5947 + }, + { + "epoch": 82.61484716157206, + "grad_norm": 0.016706543043255806, + "learning_rate": 0.0006, + "loss": 4.229398250579834, + "step": 5948 + }, + { + "epoch": 82.62882096069869, + "grad_norm": 0.015630796551704407, + "learning_rate": 0.0006, + "loss": 4.181282997131348, + "step": 5949 + }, + { + "epoch": 82.64279475982532, + "grad_norm": 0.013103844597935677, + "learning_rate": 0.0006, + "loss": 4.309617042541504, + "step": 5950 + }, + { + "epoch": 82.65676855895197, + "grad_norm": 0.016136541962623596, + "learning_rate": 0.0006, + "loss": 4.224159240722656, + "step": 5951 + }, + { + "epoch": 82.6707423580786, + "grad_norm": 0.016160814091563225, + "learning_rate": 0.0006, + "loss": 4.267970085144043, + "step": 5952 + }, + { + "epoch": 82.68471615720524, + "grad_norm": 0.01672680303454399, + "learning_rate": 0.0006, + "loss": 4.230506896972656, + "step": 5953 + }, + { + "epoch": 82.69868995633188, + "grad_norm": 0.0159586351364851, + "learning_rate": 0.0006, + "loss": 4.159542083740234, + "step": 5954 + }, + { + "epoch": 82.71266375545852, + "grad_norm": 0.01461219135671854, + "learning_rate": 0.0006, + "loss": 4.27863073348999, + "step": 5955 + }, + { + "epoch": 82.72663755458515, + "grad_norm": 0.015088613145053387, + "learning_rate": 0.0006, + "loss": 4.272418975830078, + "step": 5956 + }, + { + "epoch": 82.74061135371178, + "grad_norm": 0.01594499684870243, + "learning_rate": 0.0006, + "loss": 4.271326065063477, + "step": 5957 + }, + { + "epoch": 82.75458515283843, + "grad_norm": 0.016648447141051292, + "learning_rate": 0.0006, + "loss": 4.141478538513184, + "step": 5958 + }, + { + "epoch": 82.76855895196506, + "grad_norm": 0.01540316641330719, + "learning_rate": 0.0006, + "loss": 4.111814498901367, + "step": 5959 + }, + { + "epoch": 82.78253275109171, + "grad_norm": 0.01634364202618599, + "learning_rate": 0.0006, + "loss": 4.3113112449646, + "step": 5960 + }, + { + "epoch": 82.79650655021834, + "grad_norm": 0.014768977649509907, + "learning_rate": 0.0006, + "loss": 4.184217929840088, + "step": 5961 + }, + { + "epoch": 82.81048034934499, + "grad_norm": 0.013453755527734756, + "learning_rate": 0.0006, + "loss": 4.247729301452637, + "step": 5962 + }, + { + "epoch": 82.82445414847162, + "grad_norm": 0.014731907285749912, + "learning_rate": 0.0006, + "loss": 4.352839469909668, + "step": 5963 + }, + { + "epoch": 82.83842794759825, + "grad_norm": 0.016047345474362373, + "learning_rate": 0.0006, + "loss": 4.257283687591553, + "step": 5964 + }, + { + "epoch": 82.8524017467249, + "grad_norm": 0.017141170799732208, + "learning_rate": 0.0006, + "loss": 4.22850227355957, + "step": 5965 + }, + { + "epoch": 82.86637554585153, + "grad_norm": 0.01370396837592125, + "learning_rate": 0.0006, + "loss": 4.1169257164001465, + "step": 5966 + }, + { + "epoch": 82.88034934497817, + "grad_norm": 0.012060081586241722, + "learning_rate": 0.0006, + "loss": 4.195242404937744, + "step": 5967 + }, + { + "epoch": 82.8943231441048, + "grad_norm": 0.015528378076851368, + "learning_rate": 0.0006, + "loss": 4.294939994812012, + "step": 5968 + }, + { + "epoch": 82.90829694323143, + "grad_norm": 0.018151041120290756, + "learning_rate": 0.0006, + "loss": 4.137192726135254, + "step": 5969 + }, + { + "epoch": 82.92227074235808, + "grad_norm": 0.017895735800266266, + "learning_rate": 0.0006, + "loss": 4.174036026000977, + "step": 5970 + }, + { + "epoch": 82.93624454148471, + "grad_norm": 0.015610386617481709, + "learning_rate": 0.0006, + "loss": 4.230136871337891, + "step": 5971 + }, + { + "epoch": 82.95021834061136, + "grad_norm": 0.01339928898960352, + "learning_rate": 0.0006, + "loss": 4.199074745178223, + "step": 5972 + }, + { + "epoch": 82.96419213973799, + "grad_norm": 0.015454214997589588, + "learning_rate": 0.0006, + "loss": 4.216109752655029, + "step": 5973 + }, + { + "epoch": 82.97816593886463, + "grad_norm": 0.014978132210671902, + "learning_rate": 0.0006, + "loss": 4.131511688232422, + "step": 5974 + }, + { + "epoch": 82.99213973799127, + "grad_norm": 0.015611122362315655, + "learning_rate": 0.0006, + "loss": 4.246518135070801, + "step": 5975 + }, + { + "epoch": 83.0, + "grad_norm": 0.016682764515280724, + "learning_rate": 0.0006, + "loss": 4.110680103302002, + "step": 5976 + }, + { + "epoch": 83.0, + "eval_loss": 4.641707897186279, + "eval_runtime": 57.2799, + "eval_samples_per_second": 42.633, + "eval_steps_per_second": 1.344, + "step": 5976 + }, + { + "epoch": 83.01397379912663, + "grad_norm": 0.015007982030510902, + "learning_rate": 0.0006, + "loss": 4.100912094116211, + "step": 5977 + }, + { + "epoch": 83.02794759825328, + "grad_norm": 0.017609436064958572, + "learning_rate": 0.0006, + "loss": 4.233399868011475, + "step": 5978 + }, + { + "epoch": 83.04192139737991, + "grad_norm": 0.019817272201180458, + "learning_rate": 0.0006, + "loss": 4.191173076629639, + "step": 5979 + }, + { + "epoch": 83.05589519650655, + "grad_norm": 0.017752140760421753, + "learning_rate": 0.0006, + "loss": 4.044437885284424, + "step": 5980 + }, + { + "epoch": 83.06986899563319, + "grad_norm": 0.015482760034501553, + "learning_rate": 0.0006, + "loss": 4.319931983947754, + "step": 5981 + }, + { + "epoch": 83.08384279475983, + "grad_norm": 0.015153988264501095, + "learning_rate": 0.0006, + "loss": 4.095277309417725, + "step": 5982 + }, + { + "epoch": 83.09781659388646, + "grad_norm": 0.017025282606482506, + "learning_rate": 0.0006, + "loss": 4.235110759735107, + "step": 5983 + }, + { + "epoch": 83.1117903930131, + "grad_norm": 0.01566586270928383, + "learning_rate": 0.0006, + "loss": 4.1934685707092285, + "step": 5984 + }, + { + "epoch": 83.12576419213974, + "grad_norm": 0.016419263556599617, + "learning_rate": 0.0006, + "loss": 4.234419822692871, + "step": 5985 + }, + { + "epoch": 83.13973799126637, + "grad_norm": 0.017132675275206566, + "learning_rate": 0.0006, + "loss": 4.221447944641113, + "step": 5986 + }, + { + "epoch": 83.15371179039302, + "grad_norm": 0.016138629987835884, + "learning_rate": 0.0006, + "loss": 4.199830055236816, + "step": 5987 + }, + { + "epoch": 83.16768558951965, + "grad_norm": 0.016210174188017845, + "learning_rate": 0.0006, + "loss": 4.081717491149902, + "step": 5988 + }, + { + "epoch": 83.18165938864628, + "grad_norm": 0.016055308282375336, + "learning_rate": 0.0006, + "loss": 4.247655391693115, + "step": 5989 + }, + { + "epoch": 83.19563318777293, + "grad_norm": 0.016132952645421028, + "learning_rate": 0.0006, + "loss": 4.262535572052002, + "step": 5990 + }, + { + "epoch": 83.20960698689956, + "grad_norm": 0.01779935695230961, + "learning_rate": 0.0006, + "loss": 4.153120040893555, + "step": 5991 + }, + { + "epoch": 83.2235807860262, + "grad_norm": 0.018360646441578865, + "learning_rate": 0.0006, + "loss": 4.232644081115723, + "step": 5992 + }, + { + "epoch": 83.23755458515284, + "grad_norm": 0.019629675894975662, + "learning_rate": 0.0006, + "loss": 3.9758734703063965, + "step": 5993 + }, + { + "epoch": 83.25152838427948, + "grad_norm": 0.017819996923208237, + "learning_rate": 0.0006, + "loss": 4.196932315826416, + "step": 5994 + }, + { + "epoch": 83.26550218340611, + "grad_norm": 0.01557681243866682, + "learning_rate": 0.0006, + "loss": 4.122654914855957, + "step": 5995 + }, + { + "epoch": 83.27947598253274, + "grad_norm": 0.017858853563666344, + "learning_rate": 0.0006, + "loss": 4.185330867767334, + "step": 5996 + }, + { + "epoch": 83.29344978165939, + "grad_norm": 0.018065953627228737, + "learning_rate": 0.0006, + "loss": 4.218471050262451, + "step": 5997 + }, + { + "epoch": 83.30742358078602, + "grad_norm": 0.017247939482331276, + "learning_rate": 0.0006, + "loss": 4.127076148986816, + "step": 5998 + }, + { + "epoch": 83.32139737991267, + "grad_norm": 0.017047952860593796, + "learning_rate": 0.0006, + "loss": 4.131356716156006, + "step": 5999 + }, + { + "epoch": 83.3353711790393, + "grad_norm": 0.015905629843473434, + "learning_rate": 0.0006, + "loss": 4.2239789962768555, + "step": 6000 + }, + { + "epoch": 83.34934497816595, + "grad_norm": 0.016522347927093506, + "learning_rate": 0.0006, + "loss": 4.054897785186768, + "step": 6001 + }, + { + "epoch": 83.36331877729258, + "grad_norm": 0.01622571051120758, + "learning_rate": 0.0006, + "loss": 4.284210681915283, + "step": 6002 + }, + { + "epoch": 83.37729257641921, + "grad_norm": 0.01640467159450054, + "learning_rate": 0.0006, + "loss": 4.268840789794922, + "step": 6003 + }, + { + "epoch": 83.39126637554585, + "grad_norm": 0.015698859468102455, + "learning_rate": 0.0006, + "loss": 4.179793357849121, + "step": 6004 + }, + { + "epoch": 83.40524017467249, + "grad_norm": 0.016694528982043266, + "learning_rate": 0.0006, + "loss": 4.132297515869141, + "step": 6005 + }, + { + "epoch": 83.41921397379913, + "grad_norm": 0.015345900319516659, + "learning_rate": 0.0006, + "loss": 4.090163230895996, + "step": 6006 + }, + { + "epoch": 83.43318777292576, + "grad_norm": 0.014722758904099464, + "learning_rate": 0.0006, + "loss": 4.179166793823242, + "step": 6007 + }, + { + "epoch": 83.44716157205241, + "grad_norm": 0.014884887263178825, + "learning_rate": 0.0006, + "loss": 4.137659072875977, + "step": 6008 + }, + { + "epoch": 83.46113537117904, + "grad_norm": 0.014041902497410774, + "learning_rate": 0.0006, + "loss": 4.1202073097229, + "step": 6009 + }, + { + "epoch": 83.47510917030567, + "grad_norm": 0.012799333781003952, + "learning_rate": 0.0006, + "loss": 4.092607498168945, + "step": 6010 + }, + { + "epoch": 83.48908296943232, + "grad_norm": 0.013272707350552082, + "learning_rate": 0.0006, + "loss": 4.132715225219727, + "step": 6011 + }, + { + "epoch": 83.50305676855895, + "grad_norm": 0.012934371829032898, + "learning_rate": 0.0006, + "loss": 4.230278968811035, + "step": 6012 + }, + { + "epoch": 83.5170305676856, + "grad_norm": 0.0120006762444973, + "learning_rate": 0.0006, + "loss": 4.140068531036377, + "step": 6013 + }, + { + "epoch": 83.53100436681223, + "grad_norm": 0.013616700656712055, + "learning_rate": 0.0006, + "loss": 4.20107364654541, + "step": 6014 + }, + { + "epoch": 83.54497816593886, + "grad_norm": 0.013957851566374302, + "learning_rate": 0.0006, + "loss": 4.124945163726807, + "step": 6015 + }, + { + "epoch": 83.5589519650655, + "grad_norm": 0.012762055732309818, + "learning_rate": 0.0006, + "loss": 4.022052764892578, + "step": 6016 + }, + { + "epoch": 83.57292576419214, + "grad_norm": 0.013080576434731483, + "learning_rate": 0.0006, + "loss": 4.103252410888672, + "step": 6017 + }, + { + "epoch": 83.58689956331878, + "grad_norm": 0.013957500457763672, + "learning_rate": 0.0006, + "loss": 4.235089302062988, + "step": 6018 + }, + { + "epoch": 83.60087336244541, + "grad_norm": 0.014707164838910103, + "learning_rate": 0.0006, + "loss": 4.191570281982422, + "step": 6019 + }, + { + "epoch": 83.61484716157206, + "grad_norm": 0.012807869352400303, + "learning_rate": 0.0006, + "loss": 4.294543266296387, + "step": 6020 + }, + { + "epoch": 83.62882096069869, + "grad_norm": 0.012027994729578495, + "learning_rate": 0.0006, + "loss": 4.13849401473999, + "step": 6021 + }, + { + "epoch": 83.64279475982532, + "grad_norm": 0.012147724628448486, + "learning_rate": 0.0006, + "loss": 4.127481937408447, + "step": 6022 + }, + { + "epoch": 83.65676855895197, + "grad_norm": 0.013241392560303211, + "learning_rate": 0.0006, + "loss": 4.218972682952881, + "step": 6023 + }, + { + "epoch": 83.6707423580786, + "grad_norm": 0.014071648940443993, + "learning_rate": 0.0006, + "loss": 4.096724033355713, + "step": 6024 + }, + { + "epoch": 83.68471615720524, + "grad_norm": 0.015212797559797764, + "learning_rate": 0.0006, + "loss": 4.27097225189209, + "step": 6025 + }, + { + "epoch": 83.69868995633188, + "grad_norm": 0.017647385597229004, + "learning_rate": 0.0006, + "loss": 4.209835529327393, + "step": 6026 + }, + { + "epoch": 83.71266375545852, + "grad_norm": 0.01709647662937641, + "learning_rate": 0.0006, + "loss": 4.042106628417969, + "step": 6027 + }, + { + "epoch": 83.72663755458515, + "grad_norm": 0.014530927874147892, + "learning_rate": 0.0006, + "loss": 4.181731700897217, + "step": 6028 + }, + { + "epoch": 83.74061135371178, + "grad_norm": 0.015356711111962795, + "learning_rate": 0.0006, + "loss": 4.314183235168457, + "step": 6029 + }, + { + "epoch": 83.75458515283843, + "grad_norm": 0.015500029549002647, + "learning_rate": 0.0006, + "loss": 4.06776762008667, + "step": 6030 + }, + { + "epoch": 83.76855895196506, + "grad_norm": 0.015351332724094391, + "learning_rate": 0.0006, + "loss": 4.185141563415527, + "step": 6031 + }, + { + "epoch": 83.78253275109171, + "grad_norm": 0.016678154468536377, + "learning_rate": 0.0006, + "loss": 4.179529666900635, + "step": 6032 + }, + { + "epoch": 83.79650655021834, + "grad_norm": 0.0156265702098608, + "learning_rate": 0.0006, + "loss": 4.238584041595459, + "step": 6033 + }, + { + "epoch": 83.81048034934499, + "grad_norm": 0.016418596729636192, + "learning_rate": 0.0006, + "loss": 4.1593523025512695, + "step": 6034 + }, + { + "epoch": 83.82445414847162, + "grad_norm": 0.016113586723804474, + "learning_rate": 0.0006, + "loss": 4.157470226287842, + "step": 6035 + }, + { + "epoch": 83.83842794759825, + "grad_norm": 0.014434042386710644, + "learning_rate": 0.0006, + "loss": 4.196226119995117, + "step": 6036 + }, + { + "epoch": 83.8524017467249, + "grad_norm": 0.012647042982280254, + "learning_rate": 0.0006, + "loss": 4.130589485168457, + "step": 6037 + }, + { + "epoch": 83.86637554585153, + "grad_norm": 0.014412877149879932, + "learning_rate": 0.0006, + "loss": 4.154845237731934, + "step": 6038 + }, + { + "epoch": 83.88034934497817, + "grad_norm": 0.014222027733922005, + "learning_rate": 0.0006, + "loss": 4.189724922180176, + "step": 6039 + }, + { + "epoch": 83.8943231441048, + "grad_norm": 0.013084255158901215, + "learning_rate": 0.0006, + "loss": 4.303338527679443, + "step": 6040 + }, + { + "epoch": 83.90829694323143, + "grad_norm": 0.012659232132136822, + "learning_rate": 0.0006, + "loss": 4.08192253112793, + "step": 6041 + }, + { + "epoch": 83.92227074235808, + "grad_norm": 0.01246276218444109, + "learning_rate": 0.0006, + "loss": 4.283324718475342, + "step": 6042 + }, + { + "epoch": 83.93624454148471, + "grad_norm": 0.01340118795633316, + "learning_rate": 0.0006, + "loss": 4.117908477783203, + "step": 6043 + }, + { + "epoch": 83.95021834061136, + "grad_norm": 0.014496722258627415, + "learning_rate": 0.0006, + "loss": 4.110318660736084, + "step": 6044 + }, + { + "epoch": 83.96419213973799, + "grad_norm": 0.013348298147320747, + "learning_rate": 0.0006, + "loss": 4.185139179229736, + "step": 6045 + }, + { + "epoch": 83.97816593886463, + "grad_norm": 0.013294277712702751, + "learning_rate": 0.0006, + "loss": 4.271645545959473, + "step": 6046 + }, + { + "epoch": 83.99213973799127, + "grad_norm": 0.014431594870984554, + "learning_rate": 0.0006, + "loss": 4.161383152008057, + "step": 6047 + }, + { + "epoch": 84.0, + "grad_norm": 0.016268927603960037, + "learning_rate": 0.0006, + "loss": 4.178324222564697, + "step": 6048 + }, + { + "epoch": 84.0, + "eval_loss": 4.6307549476623535, + "eval_runtime": 56.8521, + "eval_samples_per_second": 42.954, + "eval_steps_per_second": 1.354, + "step": 6048 + }, + { + "epoch": 84.01397379912663, + "grad_norm": 0.014864908531308174, + "learning_rate": 0.0006, + "loss": 4.121300220489502, + "step": 6049 + }, + { + "epoch": 84.02794759825328, + "grad_norm": 0.014863367192447186, + "learning_rate": 0.0006, + "loss": 4.1650543212890625, + "step": 6050 + }, + { + "epoch": 84.04192139737991, + "grad_norm": 0.014379646629095078, + "learning_rate": 0.0006, + "loss": 4.0201239585876465, + "step": 6051 + }, + { + "epoch": 84.05589519650655, + "grad_norm": 0.014871201477944851, + "learning_rate": 0.0006, + "loss": 4.0820698738098145, + "step": 6052 + }, + { + "epoch": 84.06986899563319, + "grad_norm": 0.013085179962217808, + "learning_rate": 0.0006, + "loss": 4.223984718322754, + "step": 6053 + }, + { + "epoch": 84.08384279475983, + "grad_norm": 0.012485949322581291, + "learning_rate": 0.0006, + "loss": 4.182206153869629, + "step": 6054 + }, + { + "epoch": 84.09781659388646, + "grad_norm": 0.013653291389346123, + "learning_rate": 0.0006, + "loss": 4.142557144165039, + "step": 6055 + }, + { + "epoch": 84.1117903930131, + "grad_norm": 0.013930532149970531, + "learning_rate": 0.0006, + "loss": 4.13972282409668, + "step": 6056 + }, + { + "epoch": 84.12576419213974, + "grad_norm": 0.01297379657626152, + "learning_rate": 0.0006, + "loss": 4.144664764404297, + "step": 6057 + }, + { + "epoch": 84.13973799126637, + "grad_norm": 0.013210492208600044, + "learning_rate": 0.0006, + "loss": 4.253764629364014, + "step": 6058 + }, + { + "epoch": 84.15371179039302, + "grad_norm": 0.014650176279246807, + "learning_rate": 0.0006, + "loss": 4.246866226196289, + "step": 6059 + }, + { + "epoch": 84.16768558951965, + "grad_norm": 0.01482292078435421, + "learning_rate": 0.0006, + "loss": 4.140187740325928, + "step": 6060 + }, + { + "epoch": 84.18165938864628, + "grad_norm": 0.01592552848160267, + "learning_rate": 0.0006, + "loss": 4.170705318450928, + "step": 6061 + }, + { + "epoch": 84.19563318777293, + "grad_norm": 0.017152156680822372, + "learning_rate": 0.0006, + "loss": 4.126092433929443, + "step": 6062 + }, + { + "epoch": 84.20960698689956, + "grad_norm": 0.015880905091762543, + "learning_rate": 0.0006, + "loss": 4.161938190460205, + "step": 6063 + }, + { + "epoch": 84.2235807860262, + "grad_norm": 0.015451471321284771, + "learning_rate": 0.0006, + "loss": 4.263461112976074, + "step": 6064 + }, + { + "epoch": 84.23755458515284, + "grad_norm": 0.014365996234118938, + "learning_rate": 0.0006, + "loss": 4.14488410949707, + "step": 6065 + }, + { + "epoch": 84.25152838427948, + "grad_norm": 0.016364749521017075, + "learning_rate": 0.0006, + "loss": 4.112489700317383, + "step": 6066 + }, + { + "epoch": 84.26550218340611, + "grad_norm": 0.01791754551231861, + "learning_rate": 0.0006, + "loss": 4.133513450622559, + "step": 6067 + }, + { + "epoch": 84.27947598253274, + "grad_norm": 0.016609208658337593, + "learning_rate": 0.0006, + "loss": 4.201829433441162, + "step": 6068 + }, + { + "epoch": 84.29344978165939, + "grad_norm": 0.017587820068001747, + "learning_rate": 0.0006, + "loss": 4.203151702880859, + "step": 6069 + }, + { + "epoch": 84.30742358078602, + "grad_norm": 0.017047742381691933, + "learning_rate": 0.0006, + "loss": 4.077008247375488, + "step": 6070 + }, + { + "epoch": 84.32139737991267, + "grad_norm": 0.01524646207690239, + "learning_rate": 0.0006, + "loss": 4.198038101196289, + "step": 6071 + }, + { + "epoch": 84.3353711790393, + "grad_norm": 0.014196421951055527, + "learning_rate": 0.0006, + "loss": 4.164245128631592, + "step": 6072 + }, + { + "epoch": 84.34934497816595, + "grad_norm": 0.016291776672005653, + "learning_rate": 0.0006, + "loss": 4.150152206420898, + "step": 6073 + }, + { + "epoch": 84.36331877729258, + "grad_norm": 0.015627920627593994, + "learning_rate": 0.0006, + "loss": 4.110006332397461, + "step": 6074 + }, + { + "epoch": 84.37729257641921, + "grad_norm": 0.013251487165689468, + "learning_rate": 0.0006, + "loss": 4.101463317871094, + "step": 6075 + }, + { + "epoch": 84.39126637554585, + "grad_norm": 0.013601605780422688, + "learning_rate": 0.0006, + "loss": 4.191105842590332, + "step": 6076 + }, + { + "epoch": 84.40524017467249, + "grad_norm": 0.014878741465508938, + "learning_rate": 0.0006, + "loss": 4.153046131134033, + "step": 6077 + }, + { + "epoch": 84.41921397379913, + "grad_norm": 0.014783897437155247, + "learning_rate": 0.0006, + "loss": 4.362000465393066, + "step": 6078 + }, + { + "epoch": 84.43318777292576, + "grad_norm": 0.014221332035958767, + "learning_rate": 0.0006, + "loss": 4.144364833831787, + "step": 6079 + }, + { + "epoch": 84.44716157205241, + "grad_norm": 0.014911034144461155, + "learning_rate": 0.0006, + "loss": 4.088857650756836, + "step": 6080 + }, + { + "epoch": 84.46113537117904, + "grad_norm": 0.014924770221114159, + "learning_rate": 0.0006, + "loss": 4.1677446365356445, + "step": 6081 + }, + { + "epoch": 84.47510917030567, + "grad_norm": 0.014394409023225307, + "learning_rate": 0.0006, + "loss": 4.087103843688965, + "step": 6082 + }, + { + "epoch": 84.48908296943232, + "grad_norm": 0.013926676474511623, + "learning_rate": 0.0006, + "loss": 3.9972851276397705, + "step": 6083 + }, + { + "epoch": 84.50305676855895, + "grad_norm": 0.013069954700767994, + "learning_rate": 0.0006, + "loss": 3.998237133026123, + "step": 6084 + }, + { + "epoch": 84.5170305676856, + "grad_norm": 0.012748011387884617, + "learning_rate": 0.0006, + "loss": 4.195858955383301, + "step": 6085 + }, + { + "epoch": 84.53100436681223, + "grad_norm": 0.012278077192604542, + "learning_rate": 0.0006, + "loss": 4.102165222167969, + "step": 6086 + }, + { + "epoch": 84.54497816593886, + "grad_norm": 0.014259163290262222, + "learning_rate": 0.0006, + "loss": 4.092750549316406, + "step": 6087 + }, + { + "epoch": 84.5589519650655, + "grad_norm": 0.014553683809936047, + "learning_rate": 0.0006, + "loss": 4.1412763595581055, + "step": 6088 + }, + { + "epoch": 84.57292576419214, + "grad_norm": 0.013605243526399136, + "learning_rate": 0.0006, + "loss": 4.192173480987549, + "step": 6089 + }, + { + "epoch": 84.58689956331878, + "grad_norm": 0.014259358868002892, + "learning_rate": 0.0006, + "loss": 4.356374740600586, + "step": 6090 + }, + { + "epoch": 84.60087336244541, + "grad_norm": 0.014166025444865227, + "learning_rate": 0.0006, + "loss": 4.165982246398926, + "step": 6091 + }, + { + "epoch": 84.61484716157206, + "grad_norm": 0.012198803946375847, + "learning_rate": 0.0006, + "loss": 4.287566184997559, + "step": 6092 + }, + { + "epoch": 84.62882096069869, + "grad_norm": 0.013794238679111004, + "learning_rate": 0.0006, + "loss": 4.199882984161377, + "step": 6093 + }, + { + "epoch": 84.64279475982532, + "grad_norm": 0.014757642522454262, + "learning_rate": 0.0006, + "loss": 4.116573333740234, + "step": 6094 + }, + { + "epoch": 84.65676855895197, + "grad_norm": 0.014591689221560955, + "learning_rate": 0.0006, + "loss": 4.240896224975586, + "step": 6095 + }, + { + "epoch": 84.6707423580786, + "grad_norm": 0.015217579901218414, + "learning_rate": 0.0006, + "loss": 4.257142066955566, + "step": 6096 + }, + { + "epoch": 84.68471615720524, + "grad_norm": 0.015371371991932392, + "learning_rate": 0.0006, + "loss": 4.181037425994873, + "step": 6097 + }, + { + "epoch": 84.69868995633188, + "grad_norm": 0.014412983320653439, + "learning_rate": 0.0006, + "loss": 4.2207794189453125, + "step": 6098 + }, + { + "epoch": 84.71266375545852, + "grad_norm": 0.013773671351373196, + "learning_rate": 0.0006, + "loss": 4.211123466491699, + "step": 6099 + }, + { + "epoch": 84.72663755458515, + "grad_norm": 0.012838046066462994, + "learning_rate": 0.0006, + "loss": 4.175459384918213, + "step": 6100 + }, + { + "epoch": 84.74061135371178, + "grad_norm": 0.014094929210841656, + "learning_rate": 0.0006, + "loss": 4.1724958419799805, + "step": 6101 + }, + { + "epoch": 84.75458515283843, + "grad_norm": 0.014573262073099613, + "learning_rate": 0.0006, + "loss": 4.15885066986084, + "step": 6102 + }, + { + "epoch": 84.76855895196506, + "grad_norm": 0.011964819394052029, + "learning_rate": 0.0006, + "loss": 3.9804892539978027, + "step": 6103 + }, + { + "epoch": 84.78253275109171, + "grad_norm": 0.014140539802610874, + "learning_rate": 0.0006, + "loss": 4.203708648681641, + "step": 6104 + }, + { + "epoch": 84.79650655021834, + "grad_norm": 0.016618940979242325, + "learning_rate": 0.0006, + "loss": 4.147684097290039, + "step": 6105 + }, + { + "epoch": 84.81048034934499, + "grad_norm": 0.015094385482370853, + "learning_rate": 0.0006, + "loss": 4.392223834991455, + "step": 6106 + }, + { + "epoch": 84.82445414847162, + "grad_norm": 0.014386717230081558, + "learning_rate": 0.0006, + "loss": 4.253332138061523, + "step": 6107 + }, + { + "epoch": 84.83842794759825, + "grad_norm": 0.013877755962312222, + "learning_rate": 0.0006, + "loss": 4.075056076049805, + "step": 6108 + }, + { + "epoch": 84.8524017467249, + "grad_norm": 0.014062155969440937, + "learning_rate": 0.0006, + "loss": 4.072048664093018, + "step": 6109 + }, + { + "epoch": 84.86637554585153, + "grad_norm": 0.014445619657635689, + "learning_rate": 0.0006, + "loss": 4.151832580566406, + "step": 6110 + }, + { + "epoch": 84.88034934497817, + "grad_norm": 0.015116363763809204, + "learning_rate": 0.0006, + "loss": 4.088900566101074, + "step": 6111 + }, + { + "epoch": 84.8943231441048, + "grad_norm": 0.014425474219024181, + "learning_rate": 0.0006, + "loss": 4.200483322143555, + "step": 6112 + }, + { + "epoch": 84.90829694323143, + "grad_norm": 0.01449573878198862, + "learning_rate": 0.0006, + "loss": 4.239742279052734, + "step": 6113 + }, + { + "epoch": 84.92227074235808, + "grad_norm": 0.015698261559009552, + "learning_rate": 0.0006, + "loss": 4.197391510009766, + "step": 6114 + }, + { + "epoch": 84.93624454148471, + "grad_norm": 0.014545733109116554, + "learning_rate": 0.0006, + "loss": 4.035508155822754, + "step": 6115 + }, + { + "epoch": 84.95021834061136, + "grad_norm": 0.012796483002603054, + "learning_rate": 0.0006, + "loss": 4.195248126983643, + "step": 6116 + }, + { + "epoch": 84.96419213973799, + "grad_norm": 0.01264485064893961, + "learning_rate": 0.0006, + "loss": 4.158118724822998, + "step": 6117 + }, + { + "epoch": 84.97816593886463, + "grad_norm": 0.013695276342332363, + "learning_rate": 0.0006, + "loss": 4.255288124084473, + "step": 6118 + }, + { + "epoch": 84.99213973799127, + "grad_norm": 0.013459564186632633, + "learning_rate": 0.0006, + "loss": 4.204586505889893, + "step": 6119 + }, + { + "epoch": 85.0, + "grad_norm": 0.014312652871012688, + "learning_rate": 0.0006, + "loss": 4.167879104614258, + "step": 6120 + }, + { + "epoch": 85.0, + "eval_loss": 4.608152389526367, + "eval_runtime": 57.1903, + "eval_samples_per_second": 42.7, + "eval_steps_per_second": 1.346, + "step": 6120 + }, + { + "epoch": 85.01397379912663, + "grad_norm": 0.013918368145823479, + "learning_rate": 0.0006, + "loss": 4.129991054534912, + "step": 6121 + }, + { + "epoch": 85.02794759825328, + "grad_norm": 0.01604384370148182, + "learning_rate": 0.0006, + "loss": 4.233926773071289, + "step": 6122 + }, + { + "epoch": 85.04192139737991, + "grad_norm": 0.016565652564167976, + "learning_rate": 0.0006, + "loss": 4.129604339599609, + "step": 6123 + }, + { + "epoch": 85.05589519650655, + "grad_norm": 0.0149751678109169, + "learning_rate": 0.0006, + "loss": 4.165839195251465, + "step": 6124 + }, + { + "epoch": 85.06986899563319, + "grad_norm": 0.013812461867928505, + "learning_rate": 0.0006, + "loss": 4.153792858123779, + "step": 6125 + }, + { + "epoch": 85.08384279475983, + "grad_norm": 0.014432745985686779, + "learning_rate": 0.0006, + "loss": 4.170895576477051, + "step": 6126 + }, + { + "epoch": 85.09781659388646, + "grad_norm": 0.015208868309855461, + "learning_rate": 0.0006, + "loss": 4.10639762878418, + "step": 6127 + }, + { + "epoch": 85.1117903930131, + "grad_norm": 0.013209272176027298, + "learning_rate": 0.0006, + "loss": 4.200364112854004, + "step": 6128 + }, + { + "epoch": 85.12576419213974, + "grad_norm": 0.014807002618908882, + "learning_rate": 0.0006, + "loss": 4.172666549682617, + "step": 6129 + }, + { + "epoch": 85.13973799126637, + "grad_norm": 0.015662945806980133, + "learning_rate": 0.0006, + "loss": 4.145219326019287, + "step": 6130 + }, + { + "epoch": 85.15371179039302, + "grad_norm": 0.014598497189581394, + "learning_rate": 0.0006, + "loss": 4.275618553161621, + "step": 6131 + }, + { + "epoch": 85.16768558951965, + "grad_norm": 0.013778576627373695, + "learning_rate": 0.0006, + "loss": 4.204716682434082, + "step": 6132 + }, + { + "epoch": 85.18165938864628, + "grad_norm": 0.015762265771627426, + "learning_rate": 0.0006, + "loss": 4.145524024963379, + "step": 6133 + }, + { + "epoch": 85.19563318777293, + "grad_norm": 0.015579195693135262, + "learning_rate": 0.0006, + "loss": 4.150625228881836, + "step": 6134 + }, + { + "epoch": 85.20960698689956, + "grad_norm": 0.016641216352581978, + "learning_rate": 0.0006, + "loss": 4.141819953918457, + "step": 6135 + }, + { + "epoch": 85.2235807860262, + "grad_norm": 0.01632271707057953, + "learning_rate": 0.0006, + "loss": 4.030620574951172, + "step": 6136 + }, + { + "epoch": 85.23755458515284, + "grad_norm": 0.015564918518066406, + "learning_rate": 0.0006, + "loss": 4.147753715515137, + "step": 6137 + }, + { + "epoch": 85.25152838427948, + "grad_norm": 0.0161129217594862, + "learning_rate": 0.0006, + "loss": 4.128524303436279, + "step": 6138 + }, + { + "epoch": 85.26550218340611, + "grad_norm": 0.017766550183296204, + "learning_rate": 0.0006, + "loss": 4.134856224060059, + "step": 6139 + }, + { + "epoch": 85.27947598253274, + "grad_norm": 0.01869054324924946, + "learning_rate": 0.0006, + "loss": 4.231565475463867, + "step": 6140 + }, + { + "epoch": 85.29344978165939, + "grad_norm": 0.01917010359466076, + "learning_rate": 0.0006, + "loss": 4.21175479888916, + "step": 6141 + }, + { + "epoch": 85.30742358078602, + "grad_norm": 0.018116485327482224, + "learning_rate": 0.0006, + "loss": 4.122644424438477, + "step": 6142 + }, + { + "epoch": 85.32139737991267, + "grad_norm": 0.015949107706546783, + "learning_rate": 0.0006, + "loss": 4.135458946228027, + "step": 6143 + }, + { + "epoch": 85.3353711790393, + "grad_norm": 0.014783683232963085, + "learning_rate": 0.0006, + "loss": 4.118994235992432, + "step": 6144 + }, + { + "epoch": 85.34934497816595, + "grad_norm": 0.015406692400574684, + "learning_rate": 0.0006, + "loss": 4.157268524169922, + "step": 6145 + }, + { + "epoch": 85.36331877729258, + "grad_norm": 0.013893205672502518, + "learning_rate": 0.0006, + "loss": 4.056166648864746, + "step": 6146 + }, + { + "epoch": 85.37729257641921, + "grad_norm": 0.014380923472344875, + "learning_rate": 0.0006, + "loss": 4.206905364990234, + "step": 6147 + }, + { + "epoch": 85.39126637554585, + "grad_norm": 0.015669258311390877, + "learning_rate": 0.0006, + "loss": 4.213881492614746, + "step": 6148 + }, + { + "epoch": 85.40524017467249, + "grad_norm": 0.01553428452461958, + "learning_rate": 0.0006, + "loss": 4.1814374923706055, + "step": 6149 + }, + { + "epoch": 85.41921397379913, + "grad_norm": 0.01697700284421444, + "learning_rate": 0.0006, + "loss": 4.083939552307129, + "step": 6150 + }, + { + "epoch": 85.43318777292576, + "grad_norm": 0.016639070585370064, + "learning_rate": 0.0006, + "loss": 4.279311180114746, + "step": 6151 + }, + { + "epoch": 85.44716157205241, + "grad_norm": 0.016285281628370285, + "learning_rate": 0.0006, + "loss": 4.244845390319824, + "step": 6152 + }, + { + "epoch": 85.46113537117904, + "grad_norm": 0.016071317717432976, + "learning_rate": 0.0006, + "loss": 4.089814186096191, + "step": 6153 + }, + { + "epoch": 85.47510917030567, + "grad_norm": 0.01640418916940689, + "learning_rate": 0.0006, + "loss": 4.168088912963867, + "step": 6154 + }, + { + "epoch": 85.48908296943232, + "grad_norm": 0.014781414531171322, + "learning_rate": 0.0006, + "loss": 4.196092128753662, + "step": 6155 + }, + { + "epoch": 85.50305676855895, + "grad_norm": 0.013393213972449303, + "learning_rate": 0.0006, + "loss": 4.193429946899414, + "step": 6156 + }, + { + "epoch": 85.5170305676856, + "grad_norm": 0.01467406190931797, + "learning_rate": 0.0006, + "loss": 4.216714859008789, + "step": 6157 + }, + { + "epoch": 85.53100436681223, + "grad_norm": 0.014950765296816826, + "learning_rate": 0.0006, + "loss": 4.093854904174805, + "step": 6158 + }, + { + "epoch": 85.54497816593886, + "grad_norm": 0.014655301347374916, + "learning_rate": 0.0006, + "loss": 4.09483528137207, + "step": 6159 + }, + { + "epoch": 85.5589519650655, + "grad_norm": 0.01425520796328783, + "learning_rate": 0.0006, + "loss": 4.090396881103516, + "step": 6160 + }, + { + "epoch": 85.57292576419214, + "grad_norm": 0.01403278298676014, + "learning_rate": 0.0006, + "loss": 4.088724136352539, + "step": 6161 + }, + { + "epoch": 85.58689956331878, + "grad_norm": 0.01364967692643404, + "learning_rate": 0.0006, + "loss": 4.095558166503906, + "step": 6162 + }, + { + "epoch": 85.60087336244541, + "grad_norm": 0.013085267506539822, + "learning_rate": 0.0006, + "loss": 4.163962364196777, + "step": 6163 + }, + { + "epoch": 85.61484716157206, + "grad_norm": 0.0136906448751688, + "learning_rate": 0.0006, + "loss": 4.197134017944336, + "step": 6164 + }, + { + "epoch": 85.62882096069869, + "grad_norm": 0.013472451828420162, + "learning_rate": 0.0006, + "loss": 4.1578369140625, + "step": 6165 + }, + { + "epoch": 85.64279475982532, + "grad_norm": 0.012329773046076298, + "learning_rate": 0.0006, + "loss": 4.231907844543457, + "step": 6166 + }, + { + "epoch": 85.65676855895197, + "grad_norm": 0.011977200396358967, + "learning_rate": 0.0006, + "loss": 4.1552958488464355, + "step": 6167 + }, + { + "epoch": 85.6707423580786, + "grad_norm": 0.013049956411123276, + "learning_rate": 0.0006, + "loss": 4.17852783203125, + "step": 6168 + }, + { + "epoch": 85.68471615720524, + "grad_norm": 0.011440741829574108, + "learning_rate": 0.0006, + "loss": 4.144646644592285, + "step": 6169 + }, + { + "epoch": 85.69868995633188, + "grad_norm": 0.012472936883568764, + "learning_rate": 0.0006, + "loss": 4.2306413650512695, + "step": 6170 + }, + { + "epoch": 85.71266375545852, + "grad_norm": 0.013015356846153736, + "learning_rate": 0.0006, + "loss": 4.193439960479736, + "step": 6171 + }, + { + "epoch": 85.72663755458515, + "grad_norm": 0.013832319527864456, + "learning_rate": 0.0006, + "loss": 4.18073034286499, + "step": 6172 + }, + { + "epoch": 85.74061135371178, + "grad_norm": 0.014613982290029526, + "learning_rate": 0.0006, + "loss": 4.223531246185303, + "step": 6173 + }, + { + "epoch": 85.75458515283843, + "grad_norm": 0.014209083281457424, + "learning_rate": 0.0006, + "loss": 4.142666816711426, + "step": 6174 + }, + { + "epoch": 85.76855895196506, + "grad_norm": 0.013396669179201126, + "learning_rate": 0.0006, + "loss": 4.185786247253418, + "step": 6175 + }, + { + "epoch": 85.78253275109171, + "grad_norm": 0.01364929135888815, + "learning_rate": 0.0006, + "loss": 4.125226974487305, + "step": 6176 + }, + { + "epoch": 85.79650655021834, + "grad_norm": 0.012561003677546978, + "learning_rate": 0.0006, + "loss": 4.084519386291504, + "step": 6177 + }, + { + "epoch": 85.81048034934499, + "grad_norm": 0.013508481904864311, + "learning_rate": 0.0006, + "loss": 4.193173885345459, + "step": 6178 + }, + { + "epoch": 85.82445414847162, + "grad_norm": 0.014154613018035889, + "learning_rate": 0.0006, + "loss": 4.257272720336914, + "step": 6179 + }, + { + "epoch": 85.83842794759825, + "grad_norm": 0.012478531338274479, + "learning_rate": 0.0006, + "loss": 4.161380767822266, + "step": 6180 + }, + { + "epoch": 85.8524017467249, + "grad_norm": 0.013468451797962189, + "learning_rate": 0.0006, + "loss": 4.121395111083984, + "step": 6181 + }, + { + "epoch": 85.86637554585153, + "grad_norm": 0.015155194327235222, + "learning_rate": 0.0006, + "loss": 4.242815971374512, + "step": 6182 + }, + { + "epoch": 85.88034934497817, + "grad_norm": 0.016223272308707237, + "learning_rate": 0.0006, + "loss": 4.176027297973633, + "step": 6183 + }, + { + "epoch": 85.8943231441048, + "grad_norm": 0.013741295784711838, + "learning_rate": 0.0006, + "loss": 4.131257057189941, + "step": 6184 + }, + { + "epoch": 85.90829694323143, + "grad_norm": 0.014982067979872227, + "learning_rate": 0.0006, + "loss": 4.237896919250488, + "step": 6185 + }, + { + "epoch": 85.92227074235808, + "grad_norm": 0.018197322264313698, + "learning_rate": 0.0006, + "loss": 4.154962539672852, + "step": 6186 + }, + { + "epoch": 85.93624454148471, + "grad_norm": 0.01971200481057167, + "learning_rate": 0.0006, + "loss": 4.16819429397583, + "step": 6187 + }, + { + "epoch": 85.95021834061136, + "grad_norm": 0.018008720129728317, + "learning_rate": 0.0006, + "loss": 4.146164417266846, + "step": 6188 + }, + { + "epoch": 85.96419213973799, + "grad_norm": 0.015859264880418777, + "learning_rate": 0.0006, + "loss": 4.208099365234375, + "step": 6189 + }, + { + "epoch": 85.97816593886463, + "grad_norm": 0.013846252113580704, + "learning_rate": 0.0006, + "loss": 4.1881184577941895, + "step": 6190 + }, + { + "epoch": 85.99213973799127, + "grad_norm": 0.018641939386725426, + "learning_rate": 0.0006, + "loss": 4.127632141113281, + "step": 6191 + }, + { + "epoch": 86.0, + "grad_norm": 0.02385227382183075, + "learning_rate": 0.0006, + "loss": 4.281900405883789, + "step": 6192 + }, + { + "epoch": 86.0, + "eval_loss": 4.625458717346191, + "eval_runtime": 57.0634, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6192 + }, + { + "epoch": 86.01397379912663, + "grad_norm": 0.019032124429941177, + "learning_rate": 0.0006, + "loss": 4.2163190841674805, + "step": 6193 + }, + { + "epoch": 86.02794759825328, + "grad_norm": 0.018049925565719604, + "learning_rate": 0.0006, + "loss": 4.1534037590026855, + "step": 6194 + }, + { + "epoch": 86.04192139737991, + "grad_norm": 0.019116360694169998, + "learning_rate": 0.0006, + "loss": 4.212411403656006, + "step": 6195 + }, + { + "epoch": 86.05589519650655, + "grad_norm": 0.018967336043715477, + "learning_rate": 0.0006, + "loss": 4.190918922424316, + "step": 6196 + }, + { + "epoch": 86.06986899563319, + "grad_norm": 0.01945611648261547, + "learning_rate": 0.0006, + "loss": 4.18459415435791, + "step": 6197 + }, + { + "epoch": 86.08384279475983, + "grad_norm": 0.018104076385498047, + "learning_rate": 0.0006, + "loss": 4.0942816734313965, + "step": 6198 + }, + { + "epoch": 86.09781659388646, + "grad_norm": 0.01815992221236229, + "learning_rate": 0.0006, + "loss": 4.011547565460205, + "step": 6199 + }, + { + "epoch": 86.1117903930131, + "grad_norm": 0.017574431374669075, + "learning_rate": 0.0006, + "loss": 4.051593780517578, + "step": 6200 + }, + { + "epoch": 86.12576419213974, + "grad_norm": 0.01516508124768734, + "learning_rate": 0.0006, + "loss": 4.239673614501953, + "step": 6201 + }, + { + "epoch": 86.13973799126637, + "grad_norm": 0.015689702704548836, + "learning_rate": 0.0006, + "loss": 4.1085734367370605, + "step": 6202 + }, + { + "epoch": 86.15371179039302, + "grad_norm": 0.01596900261938572, + "learning_rate": 0.0006, + "loss": 4.1509599685668945, + "step": 6203 + }, + { + "epoch": 86.16768558951965, + "grad_norm": 0.01550096832215786, + "learning_rate": 0.0006, + "loss": 4.160238265991211, + "step": 6204 + }, + { + "epoch": 86.18165938864628, + "grad_norm": 0.014601529575884342, + "learning_rate": 0.0006, + "loss": 4.029219150543213, + "step": 6205 + }, + { + "epoch": 86.19563318777293, + "grad_norm": 0.014006740413606167, + "learning_rate": 0.0006, + "loss": 4.117501258850098, + "step": 6206 + }, + { + "epoch": 86.20960698689956, + "grad_norm": 0.013191957026720047, + "learning_rate": 0.0006, + "loss": 4.252002716064453, + "step": 6207 + }, + { + "epoch": 86.2235807860262, + "grad_norm": 0.013031814247369766, + "learning_rate": 0.0006, + "loss": 4.209517478942871, + "step": 6208 + }, + { + "epoch": 86.23755458515284, + "grad_norm": 0.013605714775621891, + "learning_rate": 0.0006, + "loss": 4.176309585571289, + "step": 6209 + }, + { + "epoch": 86.25152838427948, + "grad_norm": 0.012753087095916271, + "learning_rate": 0.0006, + "loss": 4.097990036010742, + "step": 6210 + }, + { + "epoch": 86.26550218340611, + "grad_norm": 0.012940628454089165, + "learning_rate": 0.0006, + "loss": 4.085589408874512, + "step": 6211 + }, + { + "epoch": 86.27947598253274, + "grad_norm": 0.014286619611084461, + "learning_rate": 0.0006, + "loss": 4.199261665344238, + "step": 6212 + }, + { + "epoch": 86.29344978165939, + "grad_norm": 0.014643525704741478, + "learning_rate": 0.0006, + "loss": 4.23718786239624, + "step": 6213 + }, + { + "epoch": 86.30742358078602, + "grad_norm": 0.015182147733867168, + "learning_rate": 0.0006, + "loss": 4.219855308532715, + "step": 6214 + }, + { + "epoch": 86.32139737991267, + "grad_norm": 0.01530102826654911, + "learning_rate": 0.0006, + "loss": 4.190094947814941, + "step": 6215 + }, + { + "epoch": 86.3353711790393, + "grad_norm": 0.015492603182792664, + "learning_rate": 0.0006, + "loss": 4.153780460357666, + "step": 6216 + }, + { + "epoch": 86.34934497816595, + "grad_norm": 0.015339907258749008, + "learning_rate": 0.0006, + "loss": 4.182865619659424, + "step": 6217 + }, + { + "epoch": 86.36331877729258, + "grad_norm": 0.015576880425214767, + "learning_rate": 0.0006, + "loss": 4.196417331695557, + "step": 6218 + }, + { + "epoch": 86.37729257641921, + "grad_norm": 0.014805737882852554, + "learning_rate": 0.0006, + "loss": 4.265018463134766, + "step": 6219 + }, + { + "epoch": 86.39126637554585, + "grad_norm": 0.01597665622830391, + "learning_rate": 0.0006, + "loss": 4.2864837646484375, + "step": 6220 + }, + { + "epoch": 86.40524017467249, + "grad_norm": 0.01693643443286419, + "learning_rate": 0.0006, + "loss": 4.118192672729492, + "step": 6221 + }, + { + "epoch": 86.41921397379913, + "grad_norm": 0.016648951917886734, + "learning_rate": 0.0006, + "loss": 4.23712682723999, + "step": 6222 + }, + { + "epoch": 86.43318777292576, + "grad_norm": 0.014809766784310341, + "learning_rate": 0.0006, + "loss": 4.251391410827637, + "step": 6223 + }, + { + "epoch": 86.44716157205241, + "grad_norm": 0.014333373866975307, + "learning_rate": 0.0006, + "loss": 4.17243766784668, + "step": 6224 + }, + { + "epoch": 86.46113537117904, + "grad_norm": 0.013912711292505264, + "learning_rate": 0.0006, + "loss": 4.250836372375488, + "step": 6225 + }, + { + "epoch": 86.47510917030567, + "grad_norm": 0.014821341261267662, + "learning_rate": 0.0006, + "loss": 4.073995590209961, + "step": 6226 + }, + { + "epoch": 86.48908296943232, + "grad_norm": 0.016634922474622726, + "learning_rate": 0.0006, + "loss": 4.155867576599121, + "step": 6227 + }, + { + "epoch": 86.50305676855895, + "grad_norm": 0.01748695969581604, + "learning_rate": 0.0006, + "loss": 4.139344692230225, + "step": 6228 + }, + { + "epoch": 86.5170305676856, + "grad_norm": 0.01569363661110401, + "learning_rate": 0.0006, + "loss": 4.23923397064209, + "step": 6229 + }, + { + "epoch": 86.53100436681223, + "grad_norm": 0.014037121087312698, + "learning_rate": 0.0006, + "loss": 4.200334072113037, + "step": 6230 + }, + { + "epoch": 86.54497816593886, + "grad_norm": 0.01401279866695404, + "learning_rate": 0.0006, + "loss": 4.035246849060059, + "step": 6231 + }, + { + "epoch": 86.5589519650655, + "grad_norm": 0.014603731222450733, + "learning_rate": 0.0006, + "loss": 4.141385555267334, + "step": 6232 + }, + { + "epoch": 86.57292576419214, + "grad_norm": 0.015113704837858677, + "learning_rate": 0.0006, + "loss": 4.130824089050293, + "step": 6233 + }, + { + "epoch": 86.58689956331878, + "grad_norm": 0.01588696800172329, + "learning_rate": 0.0006, + "loss": 4.05495023727417, + "step": 6234 + }, + { + "epoch": 86.60087336244541, + "grad_norm": 0.01593640074133873, + "learning_rate": 0.0006, + "loss": 4.22935152053833, + "step": 6235 + }, + { + "epoch": 86.61484716157206, + "grad_norm": 0.014292445965111256, + "learning_rate": 0.0006, + "loss": 4.309839248657227, + "step": 6236 + }, + { + "epoch": 86.62882096069869, + "grad_norm": 0.015471610240638256, + "learning_rate": 0.0006, + "loss": 4.268278121948242, + "step": 6237 + }, + { + "epoch": 86.64279475982532, + "grad_norm": 0.016643544659018517, + "learning_rate": 0.0006, + "loss": 4.135109901428223, + "step": 6238 + }, + { + "epoch": 86.65676855895197, + "grad_norm": 0.014733029529452324, + "learning_rate": 0.0006, + "loss": 4.060197830200195, + "step": 6239 + }, + { + "epoch": 86.6707423580786, + "grad_norm": 0.014869216829538345, + "learning_rate": 0.0006, + "loss": 4.126917362213135, + "step": 6240 + }, + { + "epoch": 86.68471615720524, + "grad_norm": 0.01608203910291195, + "learning_rate": 0.0006, + "loss": 4.183716773986816, + "step": 6241 + }, + { + "epoch": 86.69868995633188, + "grad_norm": 0.016178175806999207, + "learning_rate": 0.0006, + "loss": 4.118875503540039, + "step": 6242 + }, + { + "epoch": 86.71266375545852, + "grad_norm": 0.014576597139239311, + "learning_rate": 0.0006, + "loss": 4.261541366577148, + "step": 6243 + }, + { + "epoch": 86.72663755458515, + "grad_norm": 0.014466420747339725, + "learning_rate": 0.0006, + "loss": 4.307284355163574, + "step": 6244 + }, + { + "epoch": 86.74061135371178, + "grad_norm": 0.014042644761502743, + "learning_rate": 0.0006, + "loss": 4.2729997634887695, + "step": 6245 + }, + { + "epoch": 86.75458515283843, + "grad_norm": 0.014589524827897549, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6246 + }, + { + "epoch": 86.76855895196506, + "grad_norm": 0.018765483051538467, + "learning_rate": 0.0006, + "loss": 4.264769554138184, + "step": 6247 + }, + { + "epoch": 86.78253275109171, + "grad_norm": 0.019397731870412827, + "learning_rate": 0.0006, + "loss": 4.183379173278809, + "step": 6248 + }, + { + "epoch": 86.79650655021834, + "grad_norm": 0.017525402829051018, + "learning_rate": 0.0006, + "loss": 4.1494245529174805, + "step": 6249 + }, + { + "epoch": 86.81048034934499, + "grad_norm": 0.016310200095176697, + "learning_rate": 0.0006, + "loss": 4.149104118347168, + "step": 6250 + }, + { + "epoch": 86.82445414847162, + "grad_norm": 0.01506423857063055, + "learning_rate": 0.0006, + "loss": 4.22091817855835, + "step": 6251 + }, + { + "epoch": 86.83842794759825, + "grad_norm": 0.014200284145772457, + "learning_rate": 0.0006, + "loss": 4.177569389343262, + "step": 6252 + }, + { + "epoch": 86.8524017467249, + "grad_norm": 0.015016552060842514, + "learning_rate": 0.0006, + "loss": 4.089859962463379, + "step": 6253 + }, + { + "epoch": 86.86637554585153, + "grad_norm": 0.016054971143603325, + "learning_rate": 0.0006, + "loss": 4.157166004180908, + "step": 6254 + }, + { + "epoch": 86.88034934497817, + "grad_norm": 0.014388521201908588, + "learning_rate": 0.0006, + "loss": 4.1437578201293945, + "step": 6255 + }, + { + "epoch": 86.8943231441048, + "grad_norm": 0.015705369412899017, + "learning_rate": 0.0006, + "loss": 4.045689582824707, + "step": 6256 + }, + { + "epoch": 86.90829694323143, + "grad_norm": 0.014192395843565464, + "learning_rate": 0.0006, + "loss": 4.098793029785156, + "step": 6257 + }, + { + "epoch": 86.92227074235808, + "grad_norm": 0.013751034624874592, + "learning_rate": 0.0006, + "loss": 3.989920139312744, + "step": 6258 + }, + { + "epoch": 86.93624454148471, + "grad_norm": 0.01582193188369274, + "learning_rate": 0.0006, + "loss": 4.283449172973633, + "step": 6259 + }, + { + "epoch": 86.95021834061136, + "grad_norm": 0.01476013008505106, + "learning_rate": 0.0006, + "loss": 4.221401691436768, + "step": 6260 + }, + { + "epoch": 86.96419213973799, + "grad_norm": 0.01346337515860796, + "learning_rate": 0.0006, + "loss": 4.05896520614624, + "step": 6261 + }, + { + "epoch": 86.97816593886463, + "grad_norm": 0.012899307534098625, + "learning_rate": 0.0006, + "loss": 4.159560203552246, + "step": 6262 + }, + { + "epoch": 86.99213973799127, + "grad_norm": 0.013150574639439583, + "learning_rate": 0.0006, + "loss": 4.217330455780029, + "step": 6263 + }, + { + "epoch": 87.0, + "grad_norm": 0.014940370805561543, + "learning_rate": 0.0006, + "loss": 4.1611833572387695, + "step": 6264 + }, + { + "epoch": 87.0, + "eval_loss": 4.666043758392334, + "eval_runtime": 56.5978, + "eval_samples_per_second": 43.147, + "eval_steps_per_second": 1.36, + "step": 6264 + }, + { + "epoch": 87.01397379912663, + "grad_norm": 0.013721180148422718, + "learning_rate": 0.0006, + "loss": 4.142454147338867, + "step": 6265 + }, + { + "epoch": 87.02794759825328, + "grad_norm": 0.014805255457758904, + "learning_rate": 0.0006, + "loss": 4.121677875518799, + "step": 6266 + }, + { + "epoch": 87.04192139737991, + "grad_norm": 0.015576514415442944, + "learning_rate": 0.0006, + "loss": 4.1188435554504395, + "step": 6267 + }, + { + "epoch": 87.05589519650655, + "grad_norm": 0.01605502888560295, + "learning_rate": 0.0006, + "loss": 4.255428314208984, + "step": 6268 + }, + { + "epoch": 87.06986899563319, + "grad_norm": 0.014933343045413494, + "learning_rate": 0.0006, + "loss": 4.122596740722656, + "step": 6269 + }, + { + "epoch": 87.08384279475983, + "grad_norm": 0.014630894176661968, + "learning_rate": 0.0006, + "loss": 4.148405075073242, + "step": 6270 + }, + { + "epoch": 87.09781659388646, + "grad_norm": 0.014211163856089115, + "learning_rate": 0.0006, + "loss": 4.058030128479004, + "step": 6271 + }, + { + "epoch": 87.1117903930131, + "grad_norm": 0.014060168527066708, + "learning_rate": 0.0006, + "loss": 4.146586894989014, + "step": 6272 + }, + { + "epoch": 87.12576419213974, + "grad_norm": 0.014740100130438805, + "learning_rate": 0.0006, + "loss": 4.077739715576172, + "step": 6273 + }, + { + "epoch": 87.13973799126637, + "grad_norm": 0.013970664702355862, + "learning_rate": 0.0006, + "loss": 4.038252830505371, + "step": 6274 + }, + { + "epoch": 87.15371179039302, + "grad_norm": 0.014399951323866844, + "learning_rate": 0.0006, + "loss": 4.143700122833252, + "step": 6275 + }, + { + "epoch": 87.16768558951965, + "grad_norm": 0.01496410183608532, + "learning_rate": 0.0006, + "loss": 4.105609893798828, + "step": 6276 + }, + { + "epoch": 87.18165938864628, + "grad_norm": 0.014287355355918407, + "learning_rate": 0.0006, + "loss": 4.173645973205566, + "step": 6277 + }, + { + "epoch": 87.19563318777293, + "grad_norm": 0.0141292130574584, + "learning_rate": 0.0006, + "loss": 4.068511486053467, + "step": 6278 + }, + { + "epoch": 87.20960698689956, + "grad_norm": 0.01393523346632719, + "learning_rate": 0.0006, + "loss": 4.060687065124512, + "step": 6279 + }, + { + "epoch": 87.2235807860262, + "grad_norm": 0.01431021187454462, + "learning_rate": 0.0006, + "loss": 4.219209671020508, + "step": 6280 + }, + { + "epoch": 87.23755458515284, + "grad_norm": 0.014856027439236641, + "learning_rate": 0.0006, + "loss": 4.146749973297119, + "step": 6281 + }, + { + "epoch": 87.25152838427948, + "grad_norm": 0.013677539303898811, + "learning_rate": 0.0006, + "loss": 3.9859797954559326, + "step": 6282 + }, + { + "epoch": 87.26550218340611, + "grad_norm": 0.014078967273235321, + "learning_rate": 0.0006, + "loss": 4.129668235778809, + "step": 6283 + }, + { + "epoch": 87.27947598253274, + "grad_norm": 0.015564394183456898, + "learning_rate": 0.0006, + "loss": 4.0007829666137695, + "step": 6284 + }, + { + "epoch": 87.29344978165939, + "grad_norm": 0.015077749267220497, + "learning_rate": 0.0006, + "loss": 4.112264156341553, + "step": 6285 + }, + { + "epoch": 87.30742358078602, + "grad_norm": 0.015341218560934067, + "learning_rate": 0.0006, + "loss": 4.112249374389648, + "step": 6286 + }, + { + "epoch": 87.32139737991267, + "grad_norm": 0.016399366781115532, + "learning_rate": 0.0006, + "loss": 4.176196098327637, + "step": 6287 + }, + { + "epoch": 87.3353711790393, + "grad_norm": 0.01599763333797455, + "learning_rate": 0.0006, + "loss": 4.163441181182861, + "step": 6288 + }, + { + "epoch": 87.34934497816595, + "grad_norm": 0.015439776703715324, + "learning_rate": 0.0006, + "loss": 4.11902379989624, + "step": 6289 + }, + { + "epoch": 87.36331877729258, + "grad_norm": 0.015909943729639053, + "learning_rate": 0.0006, + "loss": 4.16621732711792, + "step": 6290 + }, + { + "epoch": 87.37729257641921, + "grad_norm": 0.016710078343749046, + "learning_rate": 0.0006, + "loss": 4.044960975646973, + "step": 6291 + }, + { + "epoch": 87.39126637554585, + "grad_norm": 0.015883933752775192, + "learning_rate": 0.0006, + "loss": 4.065184593200684, + "step": 6292 + }, + { + "epoch": 87.40524017467249, + "grad_norm": 0.013912747614085674, + "learning_rate": 0.0006, + "loss": 4.102450370788574, + "step": 6293 + }, + { + "epoch": 87.41921397379913, + "grad_norm": 0.013170420192182064, + "learning_rate": 0.0006, + "loss": 4.234364032745361, + "step": 6294 + }, + { + "epoch": 87.43318777292576, + "grad_norm": 0.014043302275240421, + "learning_rate": 0.0006, + "loss": 4.129565238952637, + "step": 6295 + }, + { + "epoch": 87.44716157205241, + "grad_norm": 0.015010139904916286, + "learning_rate": 0.0006, + "loss": 4.166678428649902, + "step": 6296 + }, + { + "epoch": 87.46113537117904, + "grad_norm": 0.014721784740686417, + "learning_rate": 0.0006, + "loss": 4.182618618011475, + "step": 6297 + }, + { + "epoch": 87.47510917030567, + "grad_norm": 0.015713468194007874, + "learning_rate": 0.0006, + "loss": 4.103855133056641, + "step": 6298 + }, + { + "epoch": 87.48908296943232, + "grad_norm": 0.017204247415065765, + "learning_rate": 0.0006, + "loss": 4.140937328338623, + "step": 6299 + }, + { + "epoch": 87.50305676855895, + "grad_norm": 0.016856668516993523, + "learning_rate": 0.0006, + "loss": 4.147859573364258, + "step": 6300 + }, + { + "epoch": 87.5170305676856, + "grad_norm": 0.017205018550157547, + "learning_rate": 0.0006, + "loss": 4.1667327880859375, + "step": 6301 + }, + { + "epoch": 87.53100436681223, + "grad_norm": 0.014821539632976055, + "learning_rate": 0.0006, + "loss": 4.089935302734375, + "step": 6302 + }, + { + "epoch": 87.54497816593886, + "grad_norm": 0.01446484588086605, + "learning_rate": 0.0006, + "loss": 4.243657112121582, + "step": 6303 + }, + { + "epoch": 87.5589519650655, + "grad_norm": 0.014374484308063984, + "learning_rate": 0.0006, + "loss": 4.14356803894043, + "step": 6304 + }, + { + "epoch": 87.57292576419214, + "grad_norm": 0.01420109998434782, + "learning_rate": 0.0006, + "loss": 4.1273651123046875, + "step": 6305 + }, + { + "epoch": 87.58689956331878, + "grad_norm": 0.014810388907790184, + "learning_rate": 0.0006, + "loss": 4.222290515899658, + "step": 6306 + }, + { + "epoch": 87.60087336244541, + "grad_norm": 0.015826983377337456, + "learning_rate": 0.0006, + "loss": 4.137996196746826, + "step": 6307 + }, + { + "epoch": 87.61484716157206, + "grad_norm": 0.014457454904913902, + "learning_rate": 0.0006, + "loss": 4.0780744552612305, + "step": 6308 + }, + { + "epoch": 87.62882096069869, + "grad_norm": 0.014458604156970978, + "learning_rate": 0.0006, + "loss": 4.1287922859191895, + "step": 6309 + }, + { + "epoch": 87.64279475982532, + "grad_norm": 0.015282071195542812, + "learning_rate": 0.0006, + "loss": 4.189508438110352, + "step": 6310 + }, + { + "epoch": 87.65676855895197, + "grad_norm": 0.017536642029881477, + "learning_rate": 0.0006, + "loss": 4.070173263549805, + "step": 6311 + }, + { + "epoch": 87.6707423580786, + "grad_norm": 0.017303580418229103, + "learning_rate": 0.0006, + "loss": 4.119133949279785, + "step": 6312 + }, + { + "epoch": 87.68471615720524, + "grad_norm": 0.016148580238223076, + "learning_rate": 0.0006, + "loss": 3.994354486465454, + "step": 6313 + }, + { + "epoch": 87.69868995633188, + "grad_norm": 0.01706942543387413, + "learning_rate": 0.0006, + "loss": 4.183113098144531, + "step": 6314 + }, + { + "epoch": 87.71266375545852, + "grad_norm": 0.016971010714769363, + "learning_rate": 0.0006, + "loss": 4.104086875915527, + "step": 6315 + }, + { + "epoch": 87.72663755458515, + "grad_norm": 0.015516847372055054, + "learning_rate": 0.0006, + "loss": 4.2409772872924805, + "step": 6316 + }, + { + "epoch": 87.74061135371178, + "grad_norm": 0.015352269634604454, + "learning_rate": 0.0006, + "loss": 4.026427268981934, + "step": 6317 + }, + { + "epoch": 87.75458515283843, + "grad_norm": 0.017702415585517883, + "learning_rate": 0.0006, + "loss": 4.177840232849121, + "step": 6318 + }, + { + "epoch": 87.76855895196506, + "grad_norm": 0.01629541628062725, + "learning_rate": 0.0006, + "loss": 4.229039192199707, + "step": 6319 + }, + { + "epoch": 87.78253275109171, + "grad_norm": 0.01503711100667715, + "learning_rate": 0.0006, + "loss": 4.214350700378418, + "step": 6320 + }, + { + "epoch": 87.79650655021834, + "grad_norm": 0.014913883060216904, + "learning_rate": 0.0006, + "loss": 4.168880939483643, + "step": 6321 + }, + { + "epoch": 87.81048034934499, + "grad_norm": 0.014513909816741943, + "learning_rate": 0.0006, + "loss": 4.162267684936523, + "step": 6322 + }, + { + "epoch": 87.82445414847162, + "grad_norm": 0.014635873027145863, + "learning_rate": 0.0006, + "loss": 4.220067977905273, + "step": 6323 + }, + { + "epoch": 87.83842794759825, + "grad_norm": 0.017102796584367752, + "learning_rate": 0.0006, + "loss": 4.1493377685546875, + "step": 6324 + }, + { + "epoch": 87.8524017467249, + "grad_norm": 0.015533884055912495, + "learning_rate": 0.0006, + "loss": 4.0646209716796875, + "step": 6325 + }, + { + "epoch": 87.86637554585153, + "grad_norm": 0.014617550186812878, + "learning_rate": 0.0006, + "loss": 4.1390533447265625, + "step": 6326 + }, + { + "epoch": 87.88034934497817, + "grad_norm": 0.014142967760562897, + "learning_rate": 0.0006, + "loss": 4.133934020996094, + "step": 6327 + }, + { + "epoch": 87.8943231441048, + "grad_norm": 0.013182433322072029, + "learning_rate": 0.0006, + "loss": 4.125327110290527, + "step": 6328 + }, + { + "epoch": 87.90829694323143, + "grad_norm": 0.013928229920566082, + "learning_rate": 0.0006, + "loss": 4.142777442932129, + "step": 6329 + }, + { + "epoch": 87.92227074235808, + "grad_norm": 0.015337013639509678, + "learning_rate": 0.0006, + "loss": 4.140210151672363, + "step": 6330 + }, + { + "epoch": 87.93624454148471, + "grad_norm": 0.014873536303639412, + "learning_rate": 0.0006, + "loss": 4.102636337280273, + "step": 6331 + }, + { + "epoch": 87.95021834061136, + "grad_norm": 0.01387431938201189, + "learning_rate": 0.0006, + "loss": 4.149059295654297, + "step": 6332 + }, + { + "epoch": 87.96419213973799, + "grad_norm": 0.013234943151473999, + "learning_rate": 0.0006, + "loss": 4.1564459800720215, + "step": 6333 + }, + { + "epoch": 87.97816593886463, + "grad_norm": 0.013781017623841763, + "learning_rate": 0.0006, + "loss": 4.189839839935303, + "step": 6334 + }, + { + "epoch": 87.99213973799127, + "grad_norm": 0.013733026571571827, + "learning_rate": 0.0006, + "loss": 4.29277229309082, + "step": 6335 + }, + { + "epoch": 88.0, + "grad_norm": 0.014752327464520931, + "learning_rate": 0.0006, + "loss": 4.163867950439453, + "step": 6336 + }, + { + "epoch": 88.0, + "eval_loss": 4.535566329956055, + "eval_runtime": 56.4677, + "eval_samples_per_second": 43.246, + "eval_steps_per_second": 1.364, + "step": 6336 + }, + { + "epoch": 88.01397379912663, + "grad_norm": 0.013958732597529888, + "learning_rate": 0.0006, + "loss": 4.200681209564209, + "step": 6337 + }, + { + "epoch": 88.02794759825328, + "grad_norm": 0.015283545479178429, + "learning_rate": 0.0006, + "loss": 4.1736016273498535, + "step": 6338 + }, + { + "epoch": 88.04192139737991, + "grad_norm": 0.016192059963941574, + "learning_rate": 0.0006, + "loss": 4.040589809417725, + "step": 6339 + }, + { + "epoch": 88.05589519650655, + "grad_norm": 0.018096154555678368, + "learning_rate": 0.0006, + "loss": 4.146696090698242, + "step": 6340 + }, + { + "epoch": 88.06986899563319, + "grad_norm": 0.01797664538025856, + "learning_rate": 0.0006, + "loss": 4.007248878479004, + "step": 6341 + }, + { + "epoch": 88.08384279475983, + "grad_norm": 0.016275612637400627, + "learning_rate": 0.0006, + "loss": 4.193206787109375, + "step": 6342 + }, + { + "epoch": 88.09781659388646, + "grad_norm": 0.014210258610546589, + "learning_rate": 0.0006, + "loss": 4.064548492431641, + "step": 6343 + }, + { + "epoch": 88.1117903930131, + "grad_norm": 0.016085151582956314, + "learning_rate": 0.0006, + "loss": 4.003668785095215, + "step": 6344 + }, + { + "epoch": 88.12576419213974, + "grad_norm": 0.01930483989417553, + "learning_rate": 0.0006, + "loss": 4.236578941345215, + "step": 6345 + }, + { + "epoch": 88.13973799126637, + "grad_norm": 0.017140906304121017, + "learning_rate": 0.0006, + "loss": 3.9593005180358887, + "step": 6346 + }, + { + "epoch": 88.15371179039302, + "grad_norm": 0.01521560549736023, + "learning_rate": 0.0006, + "loss": 4.137644290924072, + "step": 6347 + }, + { + "epoch": 88.16768558951965, + "grad_norm": 0.015788141638040543, + "learning_rate": 0.0006, + "loss": 4.124845027923584, + "step": 6348 + }, + { + "epoch": 88.18165938864628, + "grad_norm": 0.01784597896039486, + "learning_rate": 0.0006, + "loss": 4.139750957489014, + "step": 6349 + }, + { + "epoch": 88.19563318777293, + "grad_norm": 0.016266558319330215, + "learning_rate": 0.0006, + "loss": 4.074851989746094, + "step": 6350 + }, + { + "epoch": 88.20960698689956, + "grad_norm": 0.014024483971297741, + "learning_rate": 0.0006, + "loss": 4.164689064025879, + "step": 6351 + }, + { + "epoch": 88.2235807860262, + "grad_norm": 0.015694420784711838, + "learning_rate": 0.0006, + "loss": 4.219979286193848, + "step": 6352 + }, + { + "epoch": 88.23755458515284, + "grad_norm": 0.014897859655320644, + "learning_rate": 0.0006, + "loss": 4.050738334655762, + "step": 6353 + }, + { + "epoch": 88.25152838427948, + "grad_norm": 0.014827374368906021, + "learning_rate": 0.0006, + "loss": 4.137135028839111, + "step": 6354 + }, + { + "epoch": 88.26550218340611, + "grad_norm": 0.01454879529774189, + "learning_rate": 0.0006, + "loss": 3.9980897903442383, + "step": 6355 + }, + { + "epoch": 88.27947598253274, + "grad_norm": 0.015623947605490685, + "learning_rate": 0.0006, + "loss": 4.226551055908203, + "step": 6356 + }, + { + "epoch": 88.29344978165939, + "grad_norm": 0.015527035109698772, + "learning_rate": 0.0006, + "loss": 4.11053466796875, + "step": 6357 + }, + { + "epoch": 88.30742358078602, + "grad_norm": 0.015242574736475945, + "learning_rate": 0.0006, + "loss": 4.105966567993164, + "step": 6358 + }, + { + "epoch": 88.32139737991267, + "grad_norm": 0.01375538855791092, + "learning_rate": 0.0006, + "loss": 4.148453712463379, + "step": 6359 + }, + { + "epoch": 88.3353711790393, + "grad_norm": 0.01595226116478443, + "learning_rate": 0.0006, + "loss": 4.095836162567139, + "step": 6360 + }, + { + "epoch": 88.34934497816595, + "grad_norm": 0.015252740122377872, + "learning_rate": 0.0006, + "loss": 4.136335849761963, + "step": 6361 + }, + { + "epoch": 88.36331877729258, + "grad_norm": 0.014405866153538227, + "learning_rate": 0.0006, + "loss": 4.084691047668457, + "step": 6362 + }, + { + "epoch": 88.37729257641921, + "grad_norm": 0.014755490235984325, + "learning_rate": 0.0006, + "loss": 4.095475673675537, + "step": 6363 + }, + { + "epoch": 88.39126637554585, + "grad_norm": 0.015339708887040615, + "learning_rate": 0.0006, + "loss": 4.163136959075928, + "step": 6364 + }, + { + "epoch": 88.40524017467249, + "grad_norm": 0.014667164534330368, + "learning_rate": 0.0006, + "loss": 4.062265396118164, + "step": 6365 + }, + { + "epoch": 88.41921397379913, + "grad_norm": 0.014588423073291779, + "learning_rate": 0.0006, + "loss": 4.13359260559082, + "step": 6366 + }, + { + "epoch": 88.43318777292576, + "grad_norm": 0.014005377888679504, + "learning_rate": 0.0006, + "loss": 4.219711780548096, + "step": 6367 + }, + { + "epoch": 88.44716157205241, + "grad_norm": 0.012769930064678192, + "learning_rate": 0.0006, + "loss": 4.114658355712891, + "step": 6368 + }, + { + "epoch": 88.46113537117904, + "grad_norm": 0.013632693327963352, + "learning_rate": 0.0006, + "loss": 4.0675225257873535, + "step": 6369 + }, + { + "epoch": 88.47510917030567, + "grad_norm": 0.013829804956912994, + "learning_rate": 0.0006, + "loss": 4.1125969886779785, + "step": 6370 + }, + { + "epoch": 88.48908296943232, + "grad_norm": 0.013882109895348549, + "learning_rate": 0.0006, + "loss": 4.1627092361450195, + "step": 6371 + }, + { + "epoch": 88.50305676855895, + "grad_norm": 0.014023828320205212, + "learning_rate": 0.0006, + "loss": 4.01038932800293, + "step": 6372 + }, + { + "epoch": 88.5170305676856, + "grad_norm": 0.014840706251561642, + "learning_rate": 0.0006, + "loss": 4.117877960205078, + "step": 6373 + }, + { + "epoch": 88.53100436681223, + "grad_norm": 0.014165930449962616, + "learning_rate": 0.0006, + "loss": 4.1864237785339355, + "step": 6374 + }, + { + "epoch": 88.54497816593886, + "grad_norm": 0.014390201307833195, + "learning_rate": 0.0006, + "loss": 4.1044602394104, + "step": 6375 + }, + { + "epoch": 88.5589519650655, + "grad_norm": 0.015037395060062408, + "learning_rate": 0.0006, + "loss": 4.101107120513916, + "step": 6376 + }, + { + "epoch": 88.57292576419214, + "grad_norm": 0.01671452447772026, + "learning_rate": 0.0006, + "loss": 4.170581817626953, + "step": 6377 + }, + { + "epoch": 88.58689956331878, + "grad_norm": 0.01993076130747795, + "learning_rate": 0.0006, + "loss": 4.261412620544434, + "step": 6378 + }, + { + "epoch": 88.60087336244541, + "grad_norm": 0.023721016943454742, + "learning_rate": 0.0006, + "loss": 4.199363708496094, + "step": 6379 + }, + { + "epoch": 88.61484716157206, + "grad_norm": 0.023726163432002068, + "learning_rate": 0.0006, + "loss": 4.061285018920898, + "step": 6380 + }, + { + "epoch": 88.62882096069869, + "grad_norm": 0.018352877348661423, + "learning_rate": 0.0006, + "loss": 4.0721282958984375, + "step": 6381 + }, + { + "epoch": 88.64279475982532, + "grad_norm": 0.01679830253124237, + "learning_rate": 0.0006, + "loss": 4.1717939376831055, + "step": 6382 + }, + { + "epoch": 88.65676855895197, + "grad_norm": 0.017193637788295746, + "learning_rate": 0.0006, + "loss": 4.230783939361572, + "step": 6383 + }, + { + "epoch": 88.6707423580786, + "grad_norm": 0.01581464521586895, + "learning_rate": 0.0006, + "loss": 4.029906272888184, + "step": 6384 + }, + { + "epoch": 88.68471615720524, + "grad_norm": 0.014049704186618328, + "learning_rate": 0.0006, + "loss": 4.192776679992676, + "step": 6385 + }, + { + "epoch": 88.69868995633188, + "grad_norm": 0.017214450985193253, + "learning_rate": 0.0006, + "loss": 4.248916149139404, + "step": 6386 + }, + { + "epoch": 88.71266375545852, + "grad_norm": 0.0173103716224432, + "learning_rate": 0.0006, + "loss": 4.124347686767578, + "step": 6387 + }, + { + "epoch": 88.72663755458515, + "grad_norm": 0.015745092183351517, + "learning_rate": 0.0006, + "loss": 4.192878246307373, + "step": 6388 + }, + { + "epoch": 88.74061135371178, + "grad_norm": 0.013750772923231125, + "learning_rate": 0.0006, + "loss": 4.1564483642578125, + "step": 6389 + }, + { + "epoch": 88.75458515283843, + "grad_norm": 0.014969858340919018, + "learning_rate": 0.0006, + "loss": 4.0980424880981445, + "step": 6390 + }, + { + "epoch": 88.76855895196506, + "grad_norm": 0.01287173479795456, + "learning_rate": 0.0006, + "loss": 4.145712852478027, + "step": 6391 + }, + { + "epoch": 88.78253275109171, + "grad_norm": 0.012716300785541534, + "learning_rate": 0.0006, + "loss": 4.214226722717285, + "step": 6392 + }, + { + "epoch": 88.79650655021834, + "grad_norm": 0.013855204917490482, + "learning_rate": 0.0006, + "loss": 4.1667256355285645, + "step": 6393 + }, + { + "epoch": 88.81048034934499, + "grad_norm": 0.013650864362716675, + "learning_rate": 0.0006, + "loss": 4.161942481994629, + "step": 6394 + }, + { + "epoch": 88.82445414847162, + "grad_norm": 0.013780126348137856, + "learning_rate": 0.0006, + "loss": 4.087825775146484, + "step": 6395 + }, + { + "epoch": 88.83842794759825, + "grad_norm": 0.014029948972165585, + "learning_rate": 0.0006, + "loss": 4.142813205718994, + "step": 6396 + }, + { + "epoch": 88.8524017467249, + "grad_norm": 0.01322692260146141, + "learning_rate": 0.0006, + "loss": 4.106776714324951, + "step": 6397 + }, + { + "epoch": 88.86637554585153, + "grad_norm": 0.013522760011255741, + "learning_rate": 0.0006, + "loss": 4.090996265411377, + "step": 6398 + }, + { + "epoch": 88.88034934497817, + "grad_norm": 0.014474684372544289, + "learning_rate": 0.0006, + "loss": 4.124624252319336, + "step": 6399 + }, + { + "epoch": 88.8943231441048, + "grad_norm": 0.014490422792732716, + "learning_rate": 0.0006, + "loss": 4.065431594848633, + "step": 6400 + }, + { + "epoch": 88.90829694323143, + "grad_norm": 0.015379710122942924, + "learning_rate": 0.0006, + "loss": 4.207939624786377, + "step": 6401 + }, + { + "epoch": 88.92227074235808, + "grad_norm": 0.015914618968963623, + "learning_rate": 0.0006, + "loss": 4.041152000427246, + "step": 6402 + }, + { + "epoch": 88.93624454148471, + "grad_norm": 0.01480048056691885, + "learning_rate": 0.0006, + "loss": 4.138754844665527, + "step": 6403 + }, + { + "epoch": 88.95021834061136, + "grad_norm": 0.013734851032495499, + "learning_rate": 0.0006, + "loss": 4.25323486328125, + "step": 6404 + }, + { + "epoch": 88.96419213973799, + "grad_norm": 0.01452311035245657, + "learning_rate": 0.0006, + "loss": 4.115228176116943, + "step": 6405 + }, + { + "epoch": 88.97816593886463, + "grad_norm": 0.016197681427001953, + "learning_rate": 0.0006, + "loss": 4.099531173706055, + "step": 6406 + }, + { + "epoch": 88.99213973799127, + "grad_norm": 0.017169523984193802, + "learning_rate": 0.0006, + "loss": 4.182981491088867, + "step": 6407 + }, + { + "epoch": 89.0, + "grad_norm": 0.017855646088719368, + "learning_rate": 0.0006, + "loss": 4.1470136642456055, + "step": 6408 + }, + { + "epoch": 89.0, + "eval_loss": 4.6035614013671875, + "eval_runtime": 57.0625, + "eval_samples_per_second": 42.795, + "eval_steps_per_second": 1.349, + "step": 6408 + }, + { + "epoch": 89.01397379912663, + "grad_norm": 0.014834101311862469, + "learning_rate": 0.0006, + "loss": 4.159243583679199, + "step": 6409 + }, + { + "epoch": 89.02794759825328, + "grad_norm": 0.015316772274672985, + "learning_rate": 0.0006, + "loss": 4.0709123611450195, + "step": 6410 + }, + { + "epoch": 89.04192139737991, + "grad_norm": 0.01576591096818447, + "learning_rate": 0.0006, + "loss": 4.0921196937561035, + "step": 6411 + }, + { + "epoch": 89.05589519650655, + "grad_norm": 0.015948045998811722, + "learning_rate": 0.0006, + "loss": 4.181877136230469, + "step": 6412 + }, + { + "epoch": 89.06986899563319, + "grad_norm": 0.016612380743026733, + "learning_rate": 0.0006, + "loss": 4.081249237060547, + "step": 6413 + }, + { + "epoch": 89.08384279475983, + "grad_norm": 0.017455685883760452, + "learning_rate": 0.0006, + "loss": 4.175990104675293, + "step": 6414 + }, + { + "epoch": 89.09781659388646, + "grad_norm": 0.015040521509945393, + "learning_rate": 0.0006, + "loss": 4.121378421783447, + "step": 6415 + }, + { + "epoch": 89.1117903930131, + "grad_norm": 0.014513209462165833, + "learning_rate": 0.0006, + "loss": 3.996107816696167, + "step": 6416 + }, + { + "epoch": 89.12576419213974, + "grad_norm": 0.015455638989806175, + "learning_rate": 0.0006, + "loss": 4.073675155639648, + "step": 6417 + }, + { + "epoch": 89.13973799126637, + "grad_norm": 0.015468129888176918, + "learning_rate": 0.0006, + "loss": 4.130242824554443, + "step": 6418 + }, + { + "epoch": 89.15371179039302, + "grad_norm": 0.015284297987818718, + "learning_rate": 0.0006, + "loss": 4.18555212020874, + "step": 6419 + }, + { + "epoch": 89.16768558951965, + "grad_norm": 0.01476874016225338, + "learning_rate": 0.0006, + "loss": 4.134175777435303, + "step": 6420 + }, + { + "epoch": 89.18165938864628, + "grad_norm": 0.015836201608181, + "learning_rate": 0.0006, + "loss": 4.094529151916504, + "step": 6421 + }, + { + "epoch": 89.19563318777293, + "grad_norm": 0.0161928478628397, + "learning_rate": 0.0006, + "loss": 4.12647819519043, + "step": 6422 + }, + { + "epoch": 89.20960698689956, + "grad_norm": 0.015680814161896706, + "learning_rate": 0.0006, + "loss": 3.9620392322540283, + "step": 6423 + }, + { + "epoch": 89.2235807860262, + "grad_norm": 0.01584681309759617, + "learning_rate": 0.0006, + "loss": 4.049304008483887, + "step": 6424 + }, + { + "epoch": 89.23755458515284, + "grad_norm": 0.01452907919883728, + "learning_rate": 0.0006, + "loss": 4.064027786254883, + "step": 6425 + }, + { + "epoch": 89.25152838427948, + "grad_norm": 0.014002552255988121, + "learning_rate": 0.0006, + "loss": 4.151862621307373, + "step": 6426 + }, + { + "epoch": 89.26550218340611, + "grad_norm": 0.014658625237643719, + "learning_rate": 0.0006, + "loss": 4.1672210693359375, + "step": 6427 + }, + { + "epoch": 89.27947598253274, + "grad_norm": 0.015021766535937786, + "learning_rate": 0.0006, + "loss": 3.9742586612701416, + "step": 6428 + }, + { + "epoch": 89.29344978165939, + "grad_norm": 0.014681321568787098, + "learning_rate": 0.0006, + "loss": 4.024159908294678, + "step": 6429 + }, + { + "epoch": 89.30742358078602, + "grad_norm": 0.014161549508571625, + "learning_rate": 0.0006, + "loss": 4.105640888214111, + "step": 6430 + }, + { + "epoch": 89.32139737991267, + "grad_norm": 0.014801453799009323, + "learning_rate": 0.0006, + "loss": 4.123049736022949, + "step": 6431 + }, + { + "epoch": 89.3353711790393, + "grad_norm": 0.014307181350886822, + "learning_rate": 0.0006, + "loss": 4.086954593658447, + "step": 6432 + }, + { + "epoch": 89.34934497816595, + "grad_norm": 0.013299369253218174, + "learning_rate": 0.0006, + "loss": 4.109533786773682, + "step": 6433 + }, + { + "epoch": 89.36331877729258, + "grad_norm": 0.014388669282197952, + "learning_rate": 0.0006, + "loss": 4.222370147705078, + "step": 6434 + }, + { + "epoch": 89.37729257641921, + "grad_norm": 0.014501879923045635, + "learning_rate": 0.0006, + "loss": 4.224178314208984, + "step": 6435 + }, + { + "epoch": 89.39126637554585, + "grad_norm": 0.014109021984040737, + "learning_rate": 0.0006, + "loss": 4.112702369689941, + "step": 6436 + }, + { + "epoch": 89.40524017467249, + "grad_norm": 0.013521002605557442, + "learning_rate": 0.0006, + "loss": 4.010124683380127, + "step": 6437 + }, + { + "epoch": 89.41921397379913, + "grad_norm": 0.013288144953548908, + "learning_rate": 0.0006, + "loss": 4.127592086791992, + "step": 6438 + }, + { + "epoch": 89.43318777292576, + "grad_norm": 0.013882486149668694, + "learning_rate": 0.0006, + "loss": 4.031871318817139, + "step": 6439 + }, + { + "epoch": 89.44716157205241, + "grad_norm": 0.01289335172623396, + "learning_rate": 0.0006, + "loss": 4.0757551193237305, + "step": 6440 + }, + { + "epoch": 89.46113537117904, + "grad_norm": 0.01356302760541439, + "learning_rate": 0.0006, + "loss": 4.075678825378418, + "step": 6441 + }, + { + "epoch": 89.47510917030567, + "grad_norm": 0.01447434164583683, + "learning_rate": 0.0006, + "loss": 4.093453407287598, + "step": 6442 + }, + { + "epoch": 89.48908296943232, + "grad_norm": 0.014443481341004372, + "learning_rate": 0.0006, + "loss": 4.075372219085693, + "step": 6443 + }, + { + "epoch": 89.50305676855895, + "grad_norm": 0.01314218994230032, + "learning_rate": 0.0006, + "loss": 4.171533584594727, + "step": 6444 + }, + { + "epoch": 89.5170305676856, + "grad_norm": 0.01348346658051014, + "learning_rate": 0.0006, + "loss": 4.227236270904541, + "step": 6445 + }, + { + "epoch": 89.53100436681223, + "grad_norm": 0.013370679691433907, + "learning_rate": 0.0006, + "loss": 3.973062515258789, + "step": 6446 + }, + { + "epoch": 89.54497816593886, + "grad_norm": 0.01410642359405756, + "learning_rate": 0.0006, + "loss": 4.231175422668457, + "step": 6447 + }, + { + "epoch": 89.5589519650655, + "grad_norm": 0.01419337373226881, + "learning_rate": 0.0006, + "loss": 4.145868301391602, + "step": 6448 + }, + { + "epoch": 89.57292576419214, + "grad_norm": 0.013691053725779057, + "learning_rate": 0.0006, + "loss": 4.175131797790527, + "step": 6449 + }, + { + "epoch": 89.58689956331878, + "grad_norm": 0.013714492321014404, + "learning_rate": 0.0006, + "loss": 4.251893997192383, + "step": 6450 + }, + { + "epoch": 89.60087336244541, + "grad_norm": 0.01393197476863861, + "learning_rate": 0.0006, + "loss": 4.047079086303711, + "step": 6451 + }, + { + "epoch": 89.61484716157206, + "grad_norm": 0.013957309536635876, + "learning_rate": 0.0006, + "loss": 4.128530502319336, + "step": 6452 + }, + { + "epoch": 89.62882096069869, + "grad_norm": 0.013346714898943901, + "learning_rate": 0.0006, + "loss": 4.2296929359436035, + "step": 6453 + }, + { + "epoch": 89.64279475982532, + "grad_norm": 0.014243829064071178, + "learning_rate": 0.0006, + "loss": 4.128233909606934, + "step": 6454 + }, + { + "epoch": 89.65676855895197, + "grad_norm": 0.01606106385588646, + "learning_rate": 0.0006, + "loss": 4.202202320098877, + "step": 6455 + }, + { + "epoch": 89.6707423580786, + "grad_norm": 0.016074348241090775, + "learning_rate": 0.0006, + "loss": 4.215323448181152, + "step": 6456 + }, + { + "epoch": 89.68471615720524, + "grad_norm": 0.01611054316163063, + "learning_rate": 0.0006, + "loss": 4.096758842468262, + "step": 6457 + }, + { + "epoch": 89.69868995633188, + "grad_norm": 0.013788217678666115, + "learning_rate": 0.0006, + "loss": 4.098560810089111, + "step": 6458 + }, + { + "epoch": 89.71266375545852, + "grad_norm": 0.014408969320356846, + "learning_rate": 0.0006, + "loss": 4.107398509979248, + "step": 6459 + }, + { + "epoch": 89.72663755458515, + "grad_norm": 0.014765267260372639, + "learning_rate": 0.0006, + "loss": 4.057718276977539, + "step": 6460 + }, + { + "epoch": 89.74061135371178, + "grad_norm": 0.017395280301570892, + "learning_rate": 0.0006, + "loss": 4.100795269012451, + "step": 6461 + }, + { + "epoch": 89.75458515283843, + "grad_norm": 0.016821149736642838, + "learning_rate": 0.0006, + "loss": 4.153731346130371, + "step": 6462 + }, + { + "epoch": 89.76855895196506, + "grad_norm": 0.015633124858140945, + "learning_rate": 0.0006, + "loss": 4.061585903167725, + "step": 6463 + }, + { + "epoch": 89.78253275109171, + "grad_norm": 0.01451497245579958, + "learning_rate": 0.0006, + "loss": 4.1624884605407715, + "step": 6464 + }, + { + "epoch": 89.79650655021834, + "grad_norm": 0.013868392445147038, + "learning_rate": 0.0006, + "loss": 4.178536415100098, + "step": 6465 + }, + { + "epoch": 89.81048034934499, + "grad_norm": 0.014259875752031803, + "learning_rate": 0.0006, + "loss": 4.073114395141602, + "step": 6466 + }, + { + "epoch": 89.82445414847162, + "grad_norm": 0.013651594519615173, + "learning_rate": 0.0006, + "loss": 4.050265312194824, + "step": 6467 + }, + { + "epoch": 89.83842794759825, + "grad_norm": 0.014063458889722824, + "learning_rate": 0.0006, + "loss": 4.2014689445495605, + "step": 6468 + }, + { + "epoch": 89.8524017467249, + "grad_norm": 0.014179578050971031, + "learning_rate": 0.0006, + "loss": 4.110550880432129, + "step": 6469 + }, + { + "epoch": 89.86637554585153, + "grad_norm": 0.013616854324936867, + "learning_rate": 0.0006, + "loss": 4.1776580810546875, + "step": 6470 + }, + { + "epoch": 89.88034934497817, + "grad_norm": 0.014351330697536469, + "learning_rate": 0.0006, + "loss": 4.058526992797852, + "step": 6471 + }, + { + "epoch": 89.8943231441048, + "grad_norm": 0.0141244325786829, + "learning_rate": 0.0006, + "loss": 4.068673610687256, + "step": 6472 + }, + { + "epoch": 89.90829694323143, + "grad_norm": 0.01552131213247776, + "learning_rate": 0.0006, + "loss": 4.086787223815918, + "step": 6473 + }, + { + "epoch": 89.92227074235808, + "grad_norm": 0.014614894054830074, + "learning_rate": 0.0006, + "loss": 4.2190656661987305, + "step": 6474 + }, + { + "epoch": 89.93624454148471, + "grad_norm": 0.014369758777320385, + "learning_rate": 0.0006, + "loss": 4.117021560668945, + "step": 6475 + }, + { + "epoch": 89.95021834061136, + "grad_norm": 0.015352046117186546, + "learning_rate": 0.0006, + "loss": 4.167664527893066, + "step": 6476 + }, + { + "epoch": 89.96419213973799, + "grad_norm": 0.014314834959805012, + "learning_rate": 0.0006, + "loss": 4.160182952880859, + "step": 6477 + }, + { + "epoch": 89.97816593886463, + "grad_norm": 0.013511341996490955, + "learning_rate": 0.0006, + "loss": 4.096034049987793, + "step": 6478 + }, + { + "epoch": 89.99213973799127, + "grad_norm": 0.01308482326567173, + "learning_rate": 0.0006, + "loss": 4.062689781188965, + "step": 6479 + }, + { + "epoch": 90.0, + "grad_norm": 0.01442806702107191, + "learning_rate": 0.0006, + "loss": 4.1955885887146, + "step": 6480 + }, + { + "epoch": 90.0, + "eval_loss": 4.644347667694092, + "eval_runtime": 56.8677, + "eval_samples_per_second": 42.942, + "eval_steps_per_second": 1.354, + "step": 6480 + }, + { + "epoch": 90.01397379912663, + "grad_norm": 0.014839007519185543, + "learning_rate": 0.0006, + "loss": 4.171868324279785, + "step": 6481 + }, + { + "epoch": 90.02794759825328, + "grad_norm": 0.01476900465786457, + "learning_rate": 0.0006, + "loss": 4.119754791259766, + "step": 6482 + }, + { + "epoch": 90.04192139737991, + "grad_norm": 0.015179301612079144, + "learning_rate": 0.0006, + "loss": 4.06740140914917, + "step": 6483 + }, + { + "epoch": 90.05589519650655, + "grad_norm": 0.015552980825304985, + "learning_rate": 0.0006, + "loss": 4.112268447875977, + "step": 6484 + }, + { + "epoch": 90.06986899563319, + "grad_norm": 0.014930234290659428, + "learning_rate": 0.0006, + "loss": 4.071866989135742, + "step": 6485 + }, + { + "epoch": 90.08384279475983, + "grad_norm": 0.014124148525297642, + "learning_rate": 0.0006, + "loss": 4.013731956481934, + "step": 6486 + }, + { + "epoch": 90.09781659388646, + "grad_norm": 0.014204252511262894, + "learning_rate": 0.0006, + "loss": 4.133620738983154, + "step": 6487 + }, + { + "epoch": 90.1117903930131, + "grad_norm": 0.014845543541014194, + "learning_rate": 0.0006, + "loss": 4.122956275939941, + "step": 6488 + }, + { + "epoch": 90.12576419213974, + "grad_norm": 0.0146744754165411, + "learning_rate": 0.0006, + "loss": 4.087121963500977, + "step": 6489 + }, + { + "epoch": 90.13973799126637, + "grad_norm": 0.016007190570235252, + "learning_rate": 0.0006, + "loss": 3.98740816116333, + "step": 6490 + }, + { + "epoch": 90.15371179039302, + "grad_norm": 0.018072661012411118, + "learning_rate": 0.0006, + "loss": 4.1155009269714355, + "step": 6491 + }, + { + "epoch": 90.16768558951965, + "grad_norm": 0.016868166625499725, + "learning_rate": 0.0006, + "loss": 4.123727321624756, + "step": 6492 + }, + { + "epoch": 90.18165938864628, + "grad_norm": 0.017091726884245872, + "learning_rate": 0.0006, + "loss": 4.1405229568481445, + "step": 6493 + }, + { + "epoch": 90.19563318777293, + "grad_norm": 0.018329855054616928, + "learning_rate": 0.0006, + "loss": 4.093900680541992, + "step": 6494 + }, + { + "epoch": 90.20960698689956, + "grad_norm": 0.018906457349658012, + "learning_rate": 0.0006, + "loss": 4.0552849769592285, + "step": 6495 + }, + { + "epoch": 90.2235807860262, + "grad_norm": 0.017275869846343994, + "learning_rate": 0.0006, + "loss": 4.1310625076293945, + "step": 6496 + }, + { + "epoch": 90.23755458515284, + "grad_norm": 0.015029897913336754, + "learning_rate": 0.0006, + "loss": 4.100166320800781, + "step": 6497 + }, + { + "epoch": 90.25152838427948, + "grad_norm": 0.014891199767589569, + "learning_rate": 0.0006, + "loss": 4.1314802169799805, + "step": 6498 + }, + { + "epoch": 90.26550218340611, + "grad_norm": 0.014877690002322197, + "learning_rate": 0.0006, + "loss": 4.076098442077637, + "step": 6499 + }, + { + "epoch": 90.27947598253274, + "grad_norm": 0.014670864678919315, + "learning_rate": 0.0006, + "loss": 4.066137313842773, + "step": 6500 + }, + { + "epoch": 90.29344978165939, + "grad_norm": 0.01444234699010849, + "learning_rate": 0.0006, + "loss": 4.013591766357422, + "step": 6501 + }, + { + "epoch": 90.30742358078602, + "grad_norm": 0.01339263841509819, + "learning_rate": 0.0006, + "loss": 4.117959499359131, + "step": 6502 + }, + { + "epoch": 90.32139737991267, + "grad_norm": 0.014753810130059719, + "learning_rate": 0.0006, + "loss": 4.042947769165039, + "step": 6503 + }, + { + "epoch": 90.3353711790393, + "grad_norm": 0.014947964809834957, + "learning_rate": 0.0006, + "loss": 4.112810134887695, + "step": 6504 + }, + { + "epoch": 90.34934497816595, + "grad_norm": 0.015155353583395481, + "learning_rate": 0.0006, + "loss": 4.213214874267578, + "step": 6505 + }, + { + "epoch": 90.36331877729258, + "grad_norm": 0.015502216294407845, + "learning_rate": 0.0006, + "loss": 4.127619743347168, + "step": 6506 + }, + { + "epoch": 90.37729257641921, + "grad_norm": 0.015377781353890896, + "learning_rate": 0.0006, + "loss": 4.135862350463867, + "step": 6507 + }, + { + "epoch": 90.39126637554585, + "grad_norm": 0.01662006787955761, + "learning_rate": 0.0006, + "loss": 4.076877593994141, + "step": 6508 + }, + { + "epoch": 90.40524017467249, + "grad_norm": 0.016324395313858986, + "learning_rate": 0.0006, + "loss": 4.214895725250244, + "step": 6509 + }, + { + "epoch": 90.41921397379913, + "grad_norm": 0.016099024564027786, + "learning_rate": 0.0006, + "loss": 4.139717102050781, + "step": 6510 + }, + { + "epoch": 90.43318777292576, + "grad_norm": 0.014403503388166428, + "learning_rate": 0.0006, + "loss": 4.040343284606934, + "step": 6511 + }, + { + "epoch": 90.44716157205241, + "grad_norm": 0.01413416862487793, + "learning_rate": 0.0006, + "loss": 4.087702751159668, + "step": 6512 + }, + { + "epoch": 90.46113537117904, + "grad_norm": 0.01603989489376545, + "learning_rate": 0.0006, + "loss": 4.173558712005615, + "step": 6513 + }, + { + "epoch": 90.47510917030567, + "grad_norm": 0.016846206039190292, + "learning_rate": 0.0006, + "loss": 4.1955461502075195, + "step": 6514 + }, + { + "epoch": 90.48908296943232, + "grad_norm": 0.01652812585234642, + "learning_rate": 0.0006, + "loss": 4.061838150024414, + "step": 6515 + }, + { + "epoch": 90.50305676855895, + "grad_norm": 0.015100528486073017, + "learning_rate": 0.0006, + "loss": 4.041808605194092, + "step": 6516 + }, + { + "epoch": 90.5170305676856, + "grad_norm": 0.013867822475731373, + "learning_rate": 0.0006, + "loss": 4.12662410736084, + "step": 6517 + }, + { + "epoch": 90.53100436681223, + "grad_norm": 0.013684107922017574, + "learning_rate": 0.0006, + "loss": 4.1254048347473145, + "step": 6518 + }, + { + "epoch": 90.54497816593886, + "grad_norm": 0.014200013130903244, + "learning_rate": 0.0006, + "loss": 4.183216094970703, + "step": 6519 + }, + { + "epoch": 90.5589519650655, + "grad_norm": 0.013707022182643414, + "learning_rate": 0.0006, + "loss": 4.07176399230957, + "step": 6520 + }, + { + "epoch": 90.57292576419214, + "grad_norm": 0.013996582478284836, + "learning_rate": 0.0006, + "loss": 4.006575584411621, + "step": 6521 + }, + { + "epoch": 90.58689956331878, + "grad_norm": 0.014886210672557354, + "learning_rate": 0.0006, + "loss": 4.118956565856934, + "step": 6522 + }, + { + "epoch": 90.60087336244541, + "grad_norm": 0.014022685587406158, + "learning_rate": 0.0006, + "loss": 3.96986722946167, + "step": 6523 + }, + { + "epoch": 90.61484716157206, + "grad_norm": 0.014547634869813919, + "learning_rate": 0.0006, + "loss": 4.122886657714844, + "step": 6524 + }, + { + "epoch": 90.62882096069869, + "grad_norm": 0.013682851567864418, + "learning_rate": 0.0006, + "loss": 3.988292694091797, + "step": 6525 + }, + { + "epoch": 90.64279475982532, + "grad_norm": 0.014016421511769295, + "learning_rate": 0.0006, + "loss": 4.110270977020264, + "step": 6526 + }, + { + "epoch": 90.65676855895197, + "grad_norm": 0.013392233289778233, + "learning_rate": 0.0006, + "loss": 4.065059661865234, + "step": 6527 + }, + { + "epoch": 90.6707423580786, + "grad_norm": 0.012107795104384422, + "learning_rate": 0.0006, + "loss": 4.034260272979736, + "step": 6528 + }, + { + "epoch": 90.68471615720524, + "grad_norm": 0.013198604807257652, + "learning_rate": 0.0006, + "loss": 4.162083625793457, + "step": 6529 + }, + { + "epoch": 90.69868995633188, + "grad_norm": 0.012380579486489296, + "learning_rate": 0.0006, + "loss": 4.040456295013428, + "step": 6530 + }, + { + "epoch": 90.71266375545852, + "grad_norm": 0.012784135527908802, + "learning_rate": 0.0006, + "loss": 4.0823798179626465, + "step": 6531 + }, + { + "epoch": 90.72663755458515, + "grad_norm": 0.012920220382511616, + "learning_rate": 0.0006, + "loss": 4.181064128875732, + "step": 6532 + }, + { + "epoch": 90.74061135371178, + "grad_norm": 0.01338814664632082, + "learning_rate": 0.0006, + "loss": 4.111660003662109, + "step": 6533 + }, + { + "epoch": 90.75458515283843, + "grad_norm": 0.013628040440380573, + "learning_rate": 0.0006, + "loss": 4.237852573394775, + "step": 6534 + }, + { + "epoch": 90.76855895196506, + "grad_norm": 0.013384854421019554, + "learning_rate": 0.0006, + "loss": 4.121862888336182, + "step": 6535 + }, + { + "epoch": 90.78253275109171, + "grad_norm": 0.014726929366588593, + "learning_rate": 0.0006, + "loss": 4.1695990562438965, + "step": 6536 + }, + { + "epoch": 90.79650655021834, + "grad_norm": 0.0161006897687912, + "learning_rate": 0.0006, + "loss": 4.17794132232666, + "step": 6537 + }, + { + "epoch": 90.81048034934499, + "grad_norm": 0.016761884093284607, + "learning_rate": 0.0006, + "loss": 4.093764781951904, + "step": 6538 + }, + { + "epoch": 90.82445414847162, + "grad_norm": 0.017246825620532036, + "learning_rate": 0.0006, + "loss": 3.9382331371307373, + "step": 6539 + }, + { + "epoch": 90.83842794759825, + "grad_norm": 0.018141290172934532, + "learning_rate": 0.0006, + "loss": 3.973609209060669, + "step": 6540 + }, + { + "epoch": 90.8524017467249, + "grad_norm": 0.018129250034689903, + "learning_rate": 0.0006, + "loss": 4.227414131164551, + "step": 6541 + }, + { + "epoch": 90.86637554585153, + "grad_norm": 0.018234573304653168, + "learning_rate": 0.0006, + "loss": 4.151870250701904, + "step": 6542 + }, + { + "epoch": 90.88034934497817, + "grad_norm": 0.01560172438621521, + "learning_rate": 0.0006, + "loss": 4.127397537231445, + "step": 6543 + }, + { + "epoch": 90.8943231441048, + "grad_norm": 0.014367681927978992, + "learning_rate": 0.0006, + "loss": 4.17820930480957, + "step": 6544 + }, + { + "epoch": 90.90829694323143, + "grad_norm": 0.016071319580078125, + "learning_rate": 0.0006, + "loss": 4.059854507446289, + "step": 6545 + }, + { + "epoch": 90.92227074235808, + "grad_norm": 0.01589169353246689, + "learning_rate": 0.0006, + "loss": 4.282748222351074, + "step": 6546 + }, + { + "epoch": 90.93624454148471, + "grad_norm": 0.013639772310853004, + "learning_rate": 0.0006, + "loss": 4.147368907928467, + "step": 6547 + }, + { + "epoch": 90.95021834061136, + "grad_norm": 0.014248427003622055, + "learning_rate": 0.0006, + "loss": 4.240248680114746, + "step": 6548 + }, + { + "epoch": 90.96419213973799, + "grad_norm": 0.014600320719182491, + "learning_rate": 0.0006, + "loss": 4.122123718261719, + "step": 6549 + }, + { + "epoch": 90.97816593886463, + "grad_norm": 0.014371675439178944, + "learning_rate": 0.0006, + "loss": 4.013302803039551, + "step": 6550 + }, + { + "epoch": 90.99213973799127, + "grad_norm": 0.013133526779711246, + "learning_rate": 0.0006, + "loss": 4.097136497497559, + "step": 6551 + }, + { + "epoch": 91.0, + "grad_norm": 0.015599353238940239, + "learning_rate": 0.0006, + "loss": 4.198554039001465, + "step": 6552 + }, + { + "epoch": 91.0, + "eval_loss": 4.603907585144043, + "eval_runtime": 56.9418, + "eval_samples_per_second": 42.886, + "eval_steps_per_second": 1.352, + "step": 6552 + }, + { + "epoch": 91.01397379912663, + "grad_norm": 0.015917908400297165, + "learning_rate": 0.0006, + "loss": 4.133877754211426, + "step": 6553 + }, + { + "epoch": 91.02794759825328, + "grad_norm": 0.015485800802707672, + "learning_rate": 0.0006, + "loss": 4.111617088317871, + "step": 6554 + }, + { + "epoch": 91.04192139737991, + "grad_norm": 0.012969603762030602, + "learning_rate": 0.0006, + "loss": 4.064828395843506, + "step": 6555 + }, + { + "epoch": 91.05589519650655, + "grad_norm": 0.014482097700238228, + "learning_rate": 0.0006, + "loss": 4.101710319519043, + "step": 6556 + }, + { + "epoch": 91.06986899563319, + "grad_norm": 0.017242740839719772, + "learning_rate": 0.0006, + "loss": 4.015342712402344, + "step": 6557 + }, + { + "epoch": 91.08384279475983, + "grad_norm": 0.017100440338253975, + "learning_rate": 0.0006, + "loss": 3.9801602363586426, + "step": 6558 + }, + { + "epoch": 91.09781659388646, + "grad_norm": 0.015969131141901016, + "learning_rate": 0.0006, + "loss": 4.049958229064941, + "step": 6559 + }, + { + "epoch": 91.1117903930131, + "grad_norm": 0.016018927097320557, + "learning_rate": 0.0006, + "loss": 4.094139099121094, + "step": 6560 + }, + { + "epoch": 91.12576419213974, + "grad_norm": 0.018131926655769348, + "learning_rate": 0.0006, + "loss": 3.9499974250793457, + "step": 6561 + }, + { + "epoch": 91.13973799126637, + "grad_norm": 0.019425220787525177, + "learning_rate": 0.0006, + "loss": 3.9910552501678467, + "step": 6562 + }, + { + "epoch": 91.15371179039302, + "grad_norm": 0.01908663846552372, + "learning_rate": 0.0006, + "loss": 4.032179832458496, + "step": 6563 + }, + { + "epoch": 91.16768558951965, + "grad_norm": 0.01871207356452942, + "learning_rate": 0.0006, + "loss": 4.112380027770996, + "step": 6564 + }, + { + "epoch": 91.18165938864628, + "grad_norm": 0.020867953076958656, + "learning_rate": 0.0006, + "loss": 4.06105899810791, + "step": 6565 + }, + { + "epoch": 91.19563318777293, + "grad_norm": 0.019053412601351738, + "learning_rate": 0.0006, + "loss": 3.992504596710205, + "step": 6566 + }, + { + "epoch": 91.20960698689956, + "grad_norm": 0.016084402799606323, + "learning_rate": 0.0006, + "loss": 4.100559234619141, + "step": 6567 + }, + { + "epoch": 91.2235807860262, + "grad_norm": 0.016857590526342392, + "learning_rate": 0.0006, + "loss": 4.1567277908325195, + "step": 6568 + }, + { + "epoch": 91.23755458515284, + "grad_norm": 0.01702957972884178, + "learning_rate": 0.0006, + "loss": 4.234771728515625, + "step": 6569 + }, + { + "epoch": 91.25152838427948, + "grad_norm": 0.016869494691491127, + "learning_rate": 0.0006, + "loss": 3.9931387901306152, + "step": 6570 + }, + { + "epoch": 91.26550218340611, + "grad_norm": 0.01696581020951271, + "learning_rate": 0.0006, + "loss": 4.184908866882324, + "step": 6571 + }, + { + "epoch": 91.27947598253274, + "grad_norm": 0.014982126653194427, + "learning_rate": 0.0006, + "loss": 4.036668300628662, + "step": 6572 + }, + { + "epoch": 91.29344978165939, + "grad_norm": 0.014363301917910576, + "learning_rate": 0.0006, + "loss": 4.074010848999023, + "step": 6573 + }, + { + "epoch": 91.30742358078602, + "grad_norm": 0.015492337755858898, + "learning_rate": 0.0006, + "loss": 4.127626419067383, + "step": 6574 + }, + { + "epoch": 91.32139737991267, + "grad_norm": 0.015375050716102123, + "learning_rate": 0.0006, + "loss": 4.159229755401611, + "step": 6575 + }, + { + "epoch": 91.3353711790393, + "grad_norm": 0.015068565495312214, + "learning_rate": 0.0006, + "loss": 4.172623634338379, + "step": 6576 + }, + { + "epoch": 91.34934497816595, + "grad_norm": 0.0140256742015481, + "learning_rate": 0.0006, + "loss": 4.207065105438232, + "step": 6577 + }, + { + "epoch": 91.36331877729258, + "grad_norm": 0.014220152981579304, + "learning_rate": 0.0006, + "loss": 4.131444931030273, + "step": 6578 + }, + { + "epoch": 91.37729257641921, + "grad_norm": 0.0143149234354496, + "learning_rate": 0.0006, + "loss": 4.065218448638916, + "step": 6579 + }, + { + "epoch": 91.39126637554585, + "grad_norm": 0.013149659149348736, + "learning_rate": 0.0006, + "loss": 4.106743335723877, + "step": 6580 + }, + { + "epoch": 91.40524017467249, + "grad_norm": 0.014550330117344856, + "learning_rate": 0.0006, + "loss": 4.123722553253174, + "step": 6581 + }, + { + "epoch": 91.41921397379913, + "grad_norm": 0.014611710794270039, + "learning_rate": 0.0006, + "loss": 4.079610824584961, + "step": 6582 + }, + { + "epoch": 91.43318777292576, + "grad_norm": 0.013369769789278507, + "learning_rate": 0.0006, + "loss": 4.133187770843506, + "step": 6583 + }, + { + "epoch": 91.44716157205241, + "grad_norm": 0.013243545778095722, + "learning_rate": 0.0006, + "loss": 4.042607307434082, + "step": 6584 + }, + { + "epoch": 91.46113537117904, + "grad_norm": 0.013388266786932945, + "learning_rate": 0.0006, + "loss": 4.14863395690918, + "step": 6585 + }, + { + "epoch": 91.47510917030567, + "grad_norm": 0.014333273284137249, + "learning_rate": 0.0006, + "loss": 4.109699726104736, + "step": 6586 + }, + { + "epoch": 91.48908296943232, + "grad_norm": 0.012742757797241211, + "learning_rate": 0.0006, + "loss": 4.126582145690918, + "step": 6587 + }, + { + "epoch": 91.50305676855895, + "grad_norm": 0.012880304828286171, + "learning_rate": 0.0006, + "loss": 4.205300331115723, + "step": 6588 + }, + { + "epoch": 91.5170305676856, + "grad_norm": 0.013896043412387371, + "learning_rate": 0.0006, + "loss": 4.132510185241699, + "step": 6589 + }, + { + "epoch": 91.53100436681223, + "grad_norm": 0.014361281879246235, + "learning_rate": 0.0006, + "loss": 4.1487932205200195, + "step": 6590 + }, + { + "epoch": 91.54497816593886, + "grad_norm": 0.015137741342186928, + "learning_rate": 0.0006, + "loss": 4.1564531326293945, + "step": 6591 + }, + { + "epoch": 91.5589519650655, + "grad_norm": 0.014660682529211044, + "learning_rate": 0.0006, + "loss": 4.056064128875732, + "step": 6592 + }, + { + "epoch": 91.57292576419214, + "grad_norm": 0.014888476580381393, + "learning_rate": 0.0006, + "loss": 4.0633344650268555, + "step": 6593 + }, + { + "epoch": 91.58689956331878, + "grad_norm": 0.012162433005869389, + "learning_rate": 0.0006, + "loss": 4.026293754577637, + "step": 6594 + }, + { + "epoch": 91.60087336244541, + "grad_norm": 0.012892271392047405, + "learning_rate": 0.0006, + "loss": 4.267086029052734, + "step": 6595 + }, + { + "epoch": 91.61484716157206, + "grad_norm": 0.01450759544968605, + "learning_rate": 0.0006, + "loss": 4.149359703063965, + "step": 6596 + }, + { + "epoch": 91.62882096069869, + "grad_norm": 0.01419843453913927, + "learning_rate": 0.0006, + "loss": 4.094183921813965, + "step": 6597 + }, + { + "epoch": 91.64279475982532, + "grad_norm": 0.014401616528630257, + "learning_rate": 0.0006, + "loss": 4.125172138214111, + "step": 6598 + }, + { + "epoch": 91.65676855895197, + "grad_norm": 0.013945513404905796, + "learning_rate": 0.0006, + "loss": 4.146623134613037, + "step": 6599 + }, + { + "epoch": 91.6707423580786, + "grad_norm": 0.013516047969460487, + "learning_rate": 0.0006, + "loss": 4.160526275634766, + "step": 6600 + }, + { + "epoch": 91.68471615720524, + "grad_norm": 0.013262378051877022, + "learning_rate": 0.0006, + "loss": 4.074923038482666, + "step": 6601 + }, + { + "epoch": 91.69868995633188, + "grad_norm": 0.014365669339895248, + "learning_rate": 0.0006, + "loss": 4.118094444274902, + "step": 6602 + }, + { + "epoch": 91.71266375545852, + "grad_norm": 0.016713734716176987, + "learning_rate": 0.0006, + "loss": 4.021799087524414, + "step": 6603 + }, + { + "epoch": 91.72663755458515, + "grad_norm": 0.01988680474460125, + "learning_rate": 0.0006, + "loss": 4.238436222076416, + "step": 6604 + }, + { + "epoch": 91.74061135371178, + "grad_norm": 0.022749733179807663, + "learning_rate": 0.0006, + "loss": 4.010470390319824, + "step": 6605 + }, + { + "epoch": 91.75458515283843, + "grad_norm": 0.022533461451530457, + "learning_rate": 0.0006, + "loss": 4.152201175689697, + "step": 6606 + }, + { + "epoch": 91.76855895196506, + "grad_norm": 0.02021082676947117, + "learning_rate": 0.0006, + "loss": 4.120572090148926, + "step": 6607 + }, + { + "epoch": 91.78253275109171, + "grad_norm": 0.016298197209835052, + "learning_rate": 0.0006, + "loss": 4.115815162658691, + "step": 6608 + }, + { + "epoch": 91.79650655021834, + "grad_norm": 0.017562786117196083, + "learning_rate": 0.0006, + "loss": 4.109724044799805, + "step": 6609 + }, + { + "epoch": 91.81048034934499, + "grad_norm": 0.018153153359889984, + "learning_rate": 0.0006, + "loss": 4.143759727478027, + "step": 6610 + }, + { + "epoch": 91.82445414847162, + "grad_norm": 0.014762306585907936, + "learning_rate": 0.0006, + "loss": 4.25596809387207, + "step": 6611 + }, + { + "epoch": 91.83842794759825, + "grad_norm": 0.015074343420565128, + "learning_rate": 0.0006, + "loss": 4.207723140716553, + "step": 6612 + }, + { + "epoch": 91.8524017467249, + "grad_norm": 0.015028579160571098, + "learning_rate": 0.0006, + "loss": 4.01950216293335, + "step": 6613 + }, + { + "epoch": 91.86637554585153, + "grad_norm": 0.013996795751154423, + "learning_rate": 0.0006, + "loss": 4.132997989654541, + "step": 6614 + }, + { + "epoch": 91.88034934497817, + "grad_norm": 0.014978236518800259, + "learning_rate": 0.0006, + "loss": 4.132402420043945, + "step": 6615 + }, + { + "epoch": 91.8943231441048, + "grad_norm": 0.013111517764627934, + "learning_rate": 0.0006, + "loss": 4.1223649978637695, + "step": 6616 + }, + { + "epoch": 91.90829694323143, + "grad_norm": 0.013872234150767326, + "learning_rate": 0.0006, + "loss": 4.096590042114258, + "step": 6617 + }, + { + "epoch": 91.92227074235808, + "grad_norm": 0.013642470352351665, + "learning_rate": 0.0006, + "loss": 4.236227035522461, + "step": 6618 + }, + { + "epoch": 91.93624454148471, + "grad_norm": 0.014684873633086681, + "learning_rate": 0.0006, + "loss": 4.037961959838867, + "step": 6619 + }, + { + "epoch": 91.95021834061136, + "grad_norm": 0.014498206786811352, + "learning_rate": 0.0006, + "loss": 4.043325424194336, + "step": 6620 + }, + { + "epoch": 91.96419213973799, + "grad_norm": 0.012768656946718693, + "learning_rate": 0.0006, + "loss": 4.138803482055664, + "step": 6621 + }, + { + "epoch": 91.97816593886463, + "grad_norm": 0.012155905365943909, + "learning_rate": 0.0006, + "loss": 4.163424968719482, + "step": 6622 + }, + { + "epoch": 91.99213973799127, + "grad_norm": 0.012584930285811424, + "learning_rate": 0.0006, + "loss": 4.086629867553711, + "step": 6623 + }, + { + "epoch": 92.0, + "grad_norm": 0.01578151248395443, + "learning_rate": 0.0006, + "loss": 4.108386993408203, + "step": 6624 + }, + { + "epoch": 92.0, + "eval_loss": 4.541166305541992, + "eval_runtime": 56.871, + "eval_samples_per_second": 42.939, + "eval_steps_per_second": 1.354, + "step": 6624 + }, + { + "epoch": 92.01397379912663, + "grad_norm": 0.016901545226573944, + "learning_rate": 0.0006, + "loss": 4.043694496154785, + "step": 6625 + }, + { + "epoch": 92.02794759825328, + "grad_norm": 0.016299039125442505, + "learning_rate": 0.0006, + "loss": 4.100504398345947, + "step": 6626 + }, + { + "epoch": 92.04192139737991, + "grad_norm": 0.013988777063786983, + "learning_rate": 0.0006, + "loss": 4.063943862915039, + "step": 6627 + }, + { + "epoch": 92.05589519650655, + "grad_norm": 0.01343096699565649, + "learning_rate": 0.0006, + "loss": 4.0321197509765625, + "step": 6628 + }, + { + "epoch": 92.06986899563319, + "grad_norm": 0.014343147166073322, + "learning_rate": 0.0006, + "loss": 4.095499038696289, + "step": 6629 + }, + { + "epoch": 92.08384279475983, + "grad_norm": 0.014989756047725677, + "learning_rate": 0.0006, + "loss": 4.122310638427734, + "step": 6630 + }, + { + "epoch": 92.09781659388646, + "grad_norm": 0.013929174281656742, + "learning_rate": 0.0006, + "loss": 4.02870512008667, + "step": 6631 + }, + { + "epoch": 92.1117903930131, + "grad_norm": 0.014139998704195023, + "learning_rate": 0.0006, + "loss": 4.180675506591797, + "step": 6632 + }, + { + "epoch": 92.12576419213974, + "grad_norm": 0.012920474633574486, + "learning_rate": 0.0006, + "loss": 4.111686706542969, + "step": 6633 + }, + { + "epoch": 92.13973799126637, + "grad_norm": 0.01291937381029129, + "learning_rate": 0.0006, + "loss": 3.958092212677002, + "step": 6634 + }, + { + "epoch": 92.15371179039302, + "grad_norm": 0.013250474818050861, + "learning_rate": 0.0006, + "loss": 4.075105667114258, + "step": 6635 + }, + { + "epoch": 92.16768558951965, + "grad_norm": 0.01321406476199627, + "learning_rate": 0.0006, + "loss": 4.072322368621826, + "step": 6636 + }, + { + "epoch": 92.18165938864628, + "grad_norm": 0.014002865180373192, + "learning_rate": 0.0006, + "loss": 4.061572074890137, + "step": 6637 + }, + { + "epoch": 92.19563318777293, + "grad_norm": 0.013876105658710003, + "learning_rate": 0.0006, + "loss": 4.099213600158691, + "step": 6638 + }, + { + "epoch": 92.20960698689956, + "grad_norm": 0.01464459951967001, + "learning_rate": 0.0006, + "loss": 4.1573638916015625, + "step": 6639 + }, + { + "epoch": 92.2235807860262, + "grad_norm": 0.015354706905782223, + "learning_rate": 0.0006, + "loss": 4.078714370727539, + "step": 6640 + }, + { + "epoch": 92.23755458515284, + "grad_norm": 0.015265689231455326, + "learning_rate": 0.0006, + "loss": 4.036652565002441, + "step": 6641 + }, + { + "epoch": 92.25152838427948, + "grad_norm": 0.016140855848789215, + "learning_rate": 0.0006, + "loss": 4.120541095733643, + "step": 6642 + }, + { + "epoch": 92.26550218340611, + "grad_norm": 0.017044052481651306, + "learning_rate": 0.0006, + "loss": 4.015849590301514, + "step": 6643 + }, + { + "epoch": 92.27947598253274, + "grad_norm": 0.01469327136874199, + "learning_rate": 0.0006, + "loss": 4.137286186218262, + "step": 6644 + }, + { + "epoch": 92.29344978165939, + "grad_norm": 0.014596942812204361, + "learning_rate": 0.0006, + "loss": 4.11309814453125, + "step": 6645 + }, + { + "epoch": 92.30742358078602, + "grad_norm": 0.014671559445559978, + "learning_rate": 0.0006, + "loss": 4.087950229644775, + "step": 6646 + }, + { + "epoch": 92.32139737991267, + "grad_norm": 0.015040507540106773, + "learning_rate": 0.0006, + "loss": 3.967053174972534, + "step": 6647 + }, + { + "epoch": 92.3353711790393, + "grad_norm": 0.01417640782892704, + "learning_rate": 0.0006, + "loss": 3.990346908569336, + "step": 6648 + }, + { + "epoch": 92.34934497816595, + "grad_norm": 0.013855187222361565, + "learning_rate": 0.0006, + "loss": 3.965132474899292, + "step": 6649 + }, + { + "epoch": 92.36331877729258, + "grad_norm": 0.014291216619312763, + "learning_rate": 0.0006, + "loss": 4.161717891693115, + "step": 6650 + }, + { + "epoch": 92.37729257641921, + "grad_norm": 0.014148853719234467, + "learning_rate": 0.0006, + "loss": 4.019951820373535, + "step": 6651 + }, + { + "epoch": 92.39126637554585, + "grad_norm": 0.01414303295314312, + "learning_rate": 0.0006, + "loss": 4.125976085662842, + "step": 6652 + }, + { + "epoch": 92.40524017467249, + "grad_norm": 0.01358176488429308, + "learning_rate": 0.0006, + "loss": 4.088022232055664, + "step": 6653 + }, + { + "epoch": 92.41921397379913, + "grad_norm": 0.014174354262650013, + "learning_rate": 0.0006, + "loss": 4.1137518882751465, + "step": 6654 + }, + { + "epoch": 92.43318777292576, + "grad_norm": 0.013756301254034042, + "learning_rate": 0.0006, + "loss": 4.0088725090026855, + "step": 6655 + }, + { + "epoch": 92.44716157205241, + "grad_norm": 0.014257770963013172, + "learning_rate": 0.0006, + "loss": 4.05248498916626, + "step": 6656 + }, + { + "epoch": 92.46113537117904, + "grad_norm": 0.015319934114813805, + "learning_rate": 0.0006, + "loss": 4.1375732421875, + "step": 6657 + }, + { + "epoch": 92.47510917030567, + "grad_norm": 0.014655457809567451, + "learning_rate": 0.0006, + "loss": 4.172857284545898, + "step": 6658 + }, + { + "epoch": 92.48908296943232, + "grad_norm": 0.014981928281486034, + "learning_rate": 0.0006, + "loss": 4.064846038818359, + "step": 6659 + }, + { + "epoch": 92.50305676855895, + "grad_norm": 0.015197387896478176, + "learning_rate": 0.0006, + "loss": 4.119397163391113, + "step": 6660 + }, + { + "epoch": 92.5170305676856, + "grad_norm": 0.016060935333371162, + "learning_rate": 0.0006, + "loss": 4.09212064743042, + "step": 6661 + }, + { + "epoch": 92.53100436681223, + "grad_norm": 0.014884199015796185, + "learning_rate": 0.0006, + "loss": 4.027483940124512, + "step": 6662 + }, + { + "epoch": 92.54497816593886, + "grad_norm": 0.01562194712460041, + "learning_rate": 0.0006, + "loss": 3.9722909927368164, + "step": 6663 + }, + { + "epoch": 92.5589519650655, + "grad_norm": 0.014801043085753918, + "learning_rate": 0.0006, + "loss": 4.048130989074707, + "step": 6664 + }, + { + "epoch": 92.57292576419214, + "grad_norm": 0.013954494148492813, + "learning_rate": 0.0006, + "loss": 4.050351619720459, + "step": 6665 + }, + { + "epoch": 92.58689956331878, + "grad_norm": 0.013292429968714714, + "learning_rate": 0.0006, + "loss": 4.245860576629639, + "step": 6666 + }, + { + "epoch": 92.60087336244541, + "grad_norm": 0.015737095847725868, + "learning_rate": 0.0006, + "loss": 4.154323101043701, + "step": 6667 + }, + { + "epoch": 92.61484716157206, + "grad_norm": 0.018242565914988518, + "learning_rate": 0.0006, + "loss": 4.118710517883301, + "step": 6668 + }, + { + "epoch": 92.62882096069869, + "grad_norm": 0.017666636034846306, + "learning_rate": 0.0006, + "loss": 4.1013875007629395, + "step": 6669 + }, + { + "epoch": 92.64279475982532, + "grad_norm": 0.015017223544418812, + "learning_rate": 0.0006, + "loss": 4.084001064300537, + "step": 6670 + }, + { + "epoch": 92.65676855895197, + "grad_norm": 0.01359238475561142, + "learning_rate": 0.0006, + "loss": 4.091456413269043, + "step": 6671 + }, + { + "epoch": 92.6707423580786, + "grad_norm": 0.014524881727993488, + "learning_rate": 0.0006, + "loss": 4.121245384216309, + "step": 6672 + }, + { + "epoch": 92.68471615720524, + "grad_norm": 0.016214480623602867, + "learning_rate": 0.0006, + "loss": 4.133184432983398, + "step": 6673 + }, + { + "epoch": 92.69868995633188, + "grad_norm": 0.0176643505692482, + "learning_rate": 0.0006, + "loss": 4.142827987670898, + "step": 6674 + }, + { + "epoch": 92.71266375545852, + "grad_norm": 0.017004093155264854, + "learning_rate": 0.0006, + "loss": 4.067666530609131, + "step": 6675 + }, + { + "epoch": 92.72663755458515, + "grad_norm": 0.014077397994697094, + "learning_rate": 0.0006, + "loss": 4.076465129852295, + "step": 6676 + }, + { + "epoch": 92.74061135371178, + "grad_norm": 0.015905627980828285, + "learning_rate": 0.0006, + "loss": 4.083726406097412, + "step": 6677 + }, + { + "epoch": 92.75458515283843, + "grad_norm": 0.016467567533254623, + "learning_rate": 0.0006, + "loss": 4.241217136383057, + "step": 6678 + }, + { + "epoch": 92.76855895196506, + "grad_norm": 0.01541806198656559, + "learning_rate": 0.0006, + "loss": 4.05365514755249, + "step": 6679 + }, + { + "epoch": 92.78253275109171, + "grad_norm": 0.015028242021799088, + "learning_rate": 0.0006, + "loss": 4.029950141906738, + "step": 6680 + }, + { + "epoch": 92.79650655021834, + "grad_norm": 0.014366233721375465, + "learning_rate": 0.0006, + "loss": 4.181772232055664, + "step": 6681 + }, + { + "epoch": 92.81048034934499, + "grad_norm": 0.014313179068267345, + "learning_rate": 0.0006, + "loss": 4.152658462524414, + "step": 6682 + }, + { + "epoch": 92.82445414847162, + "grad_norm": 0.016504302620887756, + "learning_rate": 0.0006, + "loss": 4.080532073974609, + "step": 6683 + }, + { + "epoch": 92.83842794759825, + "grad_norm": 0.01685677096247673, + "learning_rate": 0.0006, + "loss": 4.194268226623535, + "step": 6684 + }, + { + "epoch": 92.8524017467249, + "grad_norm": 0.018684150651097298, + "learning_rate": 0.0006, + "loss": 4.218497276306152, + "step": 6685 + }, + { + "epoch": 92.86637554585153, + "grad_norm": 0.01880732551217079, + "learning_rate": 0.0006, + "loss": 4.133942604064941, + "step": 6686 + }, + { + "epoch": 92.88034934497817, + "grad_norm": 0.016504917293787003, + "learning_rate": 0.0006, + "loss": 4.139945030212402, + "step": 6687 + }, + { + "epoch": 92.8943231441048, + "grad_norm": 0.016495356336236, + "learning_rate": 0.0006, + "loss": 4.098918914794922, + "step": 6688 + }, + { + "epoch": 92.90829694323143, + "grad_norm": 0.0170587245374918, + "learning_rate": 0.0006, + "loss": 4.025818824768066, + "step": 6689 + }, + { + "epoch": 92.92227074235808, + "grad_norm": 0.01718820631504059, + "learning_rate": 0.0006, + "loss": 4.125217437744141, + "step": 6690 + }, + { + "epoch": 92.93624454148471, + "grad_norm": 0.015789354220032692, + "learning_rate": 0.0006, + "loss": 4.150068283081055, + "step": 6691 + }, + { + "epoch": 92.95021834061136, + "grad_norm": 0.01587480492889881, + "learning_rate": 0.0006, + "loss": 4.194794654846191, + "step": 6692 + }, + { + "epoch": 92.96419213973799, + "grad_norm": 0.015669209882616997, + "learning_rate": 0.0006, + "loss": 4.0550336837768555, + "step": 6693 + }, + { + "epoch": 92.97816593886463, + "grad_norm": 0.01519355084747076, + "learning_rate": 0.0006, + "loss": 4.114232063293457, + "step": 6694 + }, + { + "epoch": 92.99213973799127, + "grad_norm": 0.014596261084079742, + "learning_rate": 0.0006, + "loss": 3.992065906524658, + "step": 6695 + }, + { + "epoch": 93.0, + "grad_norm": 0.016378438100218773, + "learning_rate": 0.0006, + "loss": 4.093693733215332, + "step": 6696 + }, + { + "epoch": 93.0, + "eval_loss": 4.541905879974365, + "eval_runtime": 57.0177, + "eval_samples_per_second": 42.829, + "eval_steps_per_second": 1.35, + "step": 6696 + }, + { + "epoch": 93.01397379912663, + "grad_norm": 0.01593143492937088, + "learning_rate": 0.0006, + "loss": 4.052624225616455, + "step": 6697 + }, + { + "epoch": 93.02794759825328, + "grad_norm": 0.016086837276816368, + "learning_rate": 0.0006, + "loss": 4.130306243896484, + "step": 6698 + }, + { + "epoch": 93.04192139737991, + "grad_norm": 0.016004323959350586, + "learning_rate": 0.0006, + "loss": 3.965893268585205, + "step": 6699 + }, + { + "epoch": 93.05589519650655, + "grad_norm": 0.017667284235358238, + "learning_rate": 0.0006, + "loss": 4.095634460449219, + "step": 6700 + }, + { + "epoch": 93.06986899563319, + "grad_norm": 0.01949942484498024, + "learning_rate": 0.0006, + "loss": 3.966041326522827, + "step": 6701 + }, + { + "epoch": 93.08384279475983, + "grad_norm": 0.021898971870541573, + "learning_rate": 0.0006, + "loss": 4.010378837585449, + "step": 6702 + }, + { + "epoch": 93.09781659388646, + "grad_norm": 0.019985264167189598, + "learning_rate": 0.0006, + "loss": 4.061577320098877, + "step": 6703 + }, + { + "epoch": 93.1117903930131, + "grad_norm": 0.016470855101943016, + "learning_rate": 0.0006, + "loss": 4.0601677894592285, + "step": 6704 + }, + { + "epoch": 93.12576419213974, + "grad_norm": 0.015600379556417465, + "learning_rate": 0.0006, + "loss": 4.0860700607299805, + "step": 6705 + }, + { + "epoch": 93.13973799126637, + "grad_norm": 0.017514333128929138, + "learning_rate": 0.0006, + "loss": 4.240778923034668, + "step": 6706 + }, + { + "epoch": 93.15371179039302, + "grad_norm": 0.01780967228114605, + "learning_rate": 0.0006, + "loss": 4.076499938964844, + "step": 6707 + }, + { + "epoch": 93.16768558951965, + "grad_norm": 0.016370438039302826, + "learning_rate": 0.0006, + "loss": 4.0677618980407715, + "step": 6708 + }, + { + "epoch": 93.18165938864628, + "grad_norm": 0.015205491334199905, + "learning_rate": 0.0006, + "loss": 4.004945755004883, + "step": 6709 + }, + { + "epoch": 93.19563318777293, + "grad_norm": 0.01537297572940588, + "learning_rate": 0.0006, + "loss": 4.121147155761719, + "step": 6710 + }, + { + "epoch": 93.20960698689956, + "grad_norm": 0.017343711107969284, + "learning_rate": 0.0006, + "loss": 4.103215217590332, + "step": 6711 + }, + { + "epoch": 93.2235807860262, + "grad_norm": 0.01618131436407566, + "learning_rate": 0.0006, + "loss": 4.083741664886475, + "step": 6712 + }, + { + "epoch": 93.23755458515284, + "grad_norm": 0.013340235687792301, + "learning_rate": 0.0006, + "loss": 4.055032730102539, + "step": 6713 + }, + { + "epoch": 93.25152838427948, + "grad_norm": 0.014084120281040668, + "learning_rate": 0.0006, + "loss": 4.005610942840576, + "step": 6714 + }, + { + "epoch": 93.26550218340611, + "grad_norm": 0.016104530543088913, + "learning_rate": 0.0006, + "loss": 4.091764450073242, + "step": 6715 + }, + { + "epoch": 93.27947598253274, + "grad_norm": 0.015232563950121403, + "learning_rate": 0.0006, + "loss": 4.105735778808594, + "step": 6716 + }, + { + "epoch": 93.29344978165939, + "grad_norm": 0.014003508724272251, + "learning_rate": 0.0006, + "loss": 4.05783224105835, + "step": 6717 + }, + { + "epoch": 93.30742358078602, + "grad_norm": 0.01614833064377308, + "learning_rate": 0.0006, + "loss": 4.124567031860352, + "step": 6718 + }, + { + "epoch": 93.32139737991267, + "grad_norm": 0.015010425820946693, + "learning_rate": 0.0006, + "loss": 4.07656192779541, + "step": 6719 + }, + { + "epoch": 93.3353711790393, + "grad_norm": 0.01444012951105833, + "learning_rate": 0.0006, + "loss": 4.122415542602539, + "step": 6720 + }, + { + "epoch": 93.34934497816595, + "grad_norm": 0.015715785324573517, + "learning_rate": 0.0006, + "loss": 4.070283889770508, + "step": 6721 + }, + { + "epoch": 93.36331877729258, + "grad_norm": 0.014976450242102146, + "learning_rate": 0.0006, + "loss": 4.087043762207031, + "step": 6722 + }, + { + "epoch": 93.37729257641921, + "grad_norm": 0.014868560247123241, + "learning_rate": 0.0006, + "loss": 3.9671802520751953, + "step": 6723 + }, + { + "epoch": 93.39126637554585, + "grad_norm": 0.013518013060092926, + "learning_rate": 0.0006, + "loss": 4.136781692504883, + "step": 6724 + }, + { + "epoch": 93.40524017467249, + "grad_norm": 0.013498986139893532, + "learning_rate": 0.0006, + "loss": 3.982309103012085, + "step": 6725 + }, + { + "epoch": 93.41921397379913, + "grad_norm": 0.013707910664379597, + "learning_rate": 0.0006, + "loss": 4.181885719299316, + "step": 6726 + }, + { + "epoch": 93.43318777292576, + "grad_norm": 0.013913207687437534, + "learning_rate": 0.0006, + "loss": 3.980070114135742, + "step": 6727 + }, + { + "epoch": 93.44716157205241, + "grad_norm": 0.013169893994927406, + "learning_rate": 0.0006, + "loss": 4.023380279541016, + "step": 6728 + }, + { + "epoch": 93.46113537117904, + "grad_norm": 0.014451324008405209, + "learning_rate": 0.0006, + "loss": 4.038372039794922, + "step": 6729 + }, + { + "epoch": 93.47510917030567, + "grad_norm": 0.014401541091501713, + "learning_rate": 0.0006, + "loss": 4.069009304046631, + "step": 6730 + }, + { + "epoch": 93.48908296943232, + "grad_norm": 0.013971361331641674, + "learning_rate": 0.0006, + "loss": 4.105923652648926, + "step": 6731 + }, + { + "epoch": 93.50305676855895, + "grad_norm": 0.013192536309361458, + "learning_rate": 0.0006, + "loss": 4.07828950881958, + "step": 6732 + }, + { + "epoch": 93.5170305676856, + "grad_norm": 0.013581741601228714, + "learning_rate": 0.0006, + "loss": 4.079203128814697, + "step": 6733 + }, + { + "epoch": 93.53100436681223, + "grad_norm": 0.01508194301277399, + "learning_rate": 0.0006, + "loss": 4.084749221801758, + "step": 6734 + }, + { + "epoch": 93.54497816593886, + "grad_norm": 0.0156723503023386, + "learning_rate": 0.0006, + "loss": 4.070216178894043, + "step": 6735 + }, + { + "epoch": 93.5589519650655, + "grad_norm": 0.014996538870036602, + "learning_rate": 0.0006, + "loss": 4.111318588256836, + "step": 6736 + }, + { + "epoch": 93.57292576419214, + "grad_norm": 0.012815269641578197, + "learning_rate": 0.0006, + "loss": 4.130414962768555, + "step": 6737 + }, + { + "epoch": 93.58689956331878, + "grad_norm": 0.012887579388916492, + "learning_rate": 0.0006, + "loss": 4.128446102142334, + "step": 6738 + }, + { + "epoch": 93.60087336244541, + "grad_norm": 0.013632671907544136, + "learning_rate": 0.0006, + "loss": 4.20977258682251, + "step": 6739 + }, + { + "epoch": 93.61484716157206, + "grad_norm": 0.014318895526230335, + "learning_rate": 0.0006, + "loss": 4.005739212036133, + "step": 6740 + }, + { + "epoch": 93.62882096069869, + "grad_norm": 0.014160127379000187, + "learning_rate": 0.0006, + "loss": 4.122544288635254, + "step": 6741 + }, + { + "epoch": 93.64279475982532, + "grad_norm": 0.01308423187583685, + "learning_rate": 0.0006, + "loss": 4.108774662017822, + "step": 6742 + }, + { + "epoch": 93.65676855895197, + "grad_norm": 0.014293737709522247, + "learning_rate": 0.0006, + "loss": 3.983156681060791, + "step": 6743 + }, + { + "epoch": 93.6707423580786, + "grad_norm": 0.015155135653913021, + "learning_rate": 0.0006, + "loss": 4.006117820739746, + "step": 6744 + }, + { + "epoch": 93.68471615720524, + "grad_norm": 0.014391839504241943, + "learning_rate": 0.0006, + "loss": 4.1658935546875, + "step": 6745 + }, + { + "epoch": 93.69868995633188, + "grad_norm": 0.014924845658242702, + "learning_rate": 0.0006, + "loss": 4.090868949890137, + "step": 6746 + }, + { + "epoch": 93.71266375545852, + "grad_norm": 0.014786627143621445, + "learning_rate": 0.0006, + "loss": 4.121716499328613, + "step": 6747 + }, + { + "epoch": 93.72663755458515, + "grad_norm": 0.016948824748396873, + "learning_rate": 0.0006, + "loss": 4.213179111480713, + "step": 6748 + }, + { + "epoch": 93.74061135371178, + "grad_norm": 0.01732620596885681, + "learning_rate": 0.0006, + "loss": 4.079215049743652, + "step": 6749 + }, + { + "epoch": 93.75458515283843, + "grad_norm": 0.016298124566674232, + "learning_rate": 0.0006, + "loss": 4.096126079559326, + "step": 6750 + }, + { + "epoch": 93.76855895196506, + "grad_norm": 0.015744732692837715, + "learning_rate": 0.0006, + "loss": 4.046300411224365, + "step": 6751 + }, + { + "epoch": 93.78253275109171, + "grad_norm": 0.016330812126398087, + "learning_rate": 0.0006, + "loss": 4.098011016845703, + "step": 6752 + }, + { + "epoch": 93.79650655021834, + "grad_norm": 0.01710600219666958, + "learning_rate": 0.0006, + "loss": 4.031402587890625, + "step": 6753 + }, + { + "epoch": 93.81048034934499, + "grad_norm": 0.01561805047094822, + "learning_rate": 0.0006, + "loss": 4.023388862609863, + "step": 6754 + }, + { + "epoch": 93.82445414847162, + "grad_norm": 0.013637243770062923, + "learning_rate": 0.0006, + "loss": 4.143238067626953, + "step": 6755 + }, + { + "epoch": 93.83842794759825, + "grad_norm": 0.013088075444102287, + "learning_rate": 0.0006, + "loss": 3.9083704948425293, + "step": 6756 + }, + { + "epoch": 93.8524017467249, + "grad_norm": 0.015893537551164627, + "learning_rate": 0.0006, + "loss": 4.102263450622559, + "step": 6757 + }, + { + "epoch": 93.86637554585153, + "grad_norm": 0.015213928185403347, + "learning_rate": 0.0006, + "loss": 4.111978530883789, + "step": 6758 + }, + { + "epoch": 93.88034934497817, + "grad_norm": 0.014543982222676277, + "learning_rate": 0.0006, + "loss": 4.05889892578125, + "step": 6759 + }, + { + "epoch": 93.8943231441048, + "grad_norm": 0.015104752965271473, + "learning_rate": 0.0006, + "loss": 4.136166572570801, + "step": 6760 + }, + { + "epoch": 93.90829694323143, + "grad_norm": 0.01505881268531084, + "learning_rate": 0.0006, + "loss": 4.061883449554443, + "step": 6761 + }, + { + "epoch": 93.92227074235808, + "grad_norm": 0.015248455107212067, + "learning_rate": 0.0006, + "loss": 4.109863758087158, + "step": 6762 + }, + { + "epoch": 93.93624454148471, + "grad_norm": 0.013651185669004917, + "learning_rate": 0.0006, + "loss": 4.061549663543701, + "step": 6763 + }, + { + "epoch": 93.95021834061136, + "grad_norm": 0.013806282542645931, + "learning_rate": 0.0006, + "loss": 4.205403804779053, + "step": 6764 + }, + { + "epoch": 93.96419213973799, + "grad_norm": 0.01538451761007309, + "learning_rate": 0.0006, + "loss": 4.14816951751709, + "step": 6765 + }, + { + "epoch": 93.97816593886463, + "grad_norm": 0.01602427288889885, + "learning_rate": 0.0006, + "loss": 4.000303268432617, + "step": 6766 + }, + { + "epoch": 93.99213973799127, + "grad_norm": 0.014558793976902962, + "learning_rate": 0.0006, + "loss": 4.096562385559082, + "step": 6767 + }, + { + "epoch": 94.0, + "grad_norm": 0.01613018475472927, + "learning_rate": 0.0006, + "loss": 4.262835502624512, + "step": 6768 + }, + { + "epoch": 94.0, + "eval_loss": 4.532126426696777, + "eval_runtime": 56.7592, + "eval_samples_per_second": 43.024, + "eval_steps_per_second": 1.357, + "step": 6768 + }, + { + "epoch": 94.01397379912663, + "grad_norm": 0.01614535041153431, + "learning_rate": 0.0006, + "loss": 4.073431491851807, + "step": 6769 + }, + { + "epoch": 94.02794759825328, + "grad_norm": 0.015986818820238113, + "learning_rate": 0.0006, + "loss": 4.070001602172852, + "step": 6770 + }, + { + "epoch": 94.04192139737991, + "grad_norm": 0.015678714960813522, + "learning_rate": 0.0006, + "loss": 4.06920051574707, + "step": 6771 + }, + { + "epoch": 94.05589519650655, + "grad_norm": 0.016593851149082184, + "learning_rate": 0.0006, + "loss": 4.095780849456787, + "step": 6772 + }, + { + "epoch": 94.06986899563319, + "grad_norm": 0.01780948042869568, + "learning_rate": 0.0006, + "loss": 4.092916488647461, + "step": 6773 + }, + { + "epoch": 94.08384279475983, + "grad_norm": 0.01755528897047043, + "learning_rate": 0.0006, + "loss": 3.9833455085754395, + "step": 6774 + }, + { + "epoch": 94.09781659388646, + "grad_norm": 0.01708311215043068, + "learning_rate": 0.0006, + "loss": 4.053336143493652, + "step": 6775 + }, + { + "epoch": 94.1117903930131, + "grad_norm": 0.017362922430038452, + "learning_rate": 0.0006, + "loss": 4.014343738555908, + "step": 6776 + }, + { + "epoch": 94.12576419213974, + "grad_norm": 0.01824064739048481, + "learning_rate": 0.0006, + "loss": 4.1215691566467285, + "step": 6777 + }, + { + "epoch": 94.13973799126637, + "grad_norm": 0.017743906006217003, + "learning_rate": 0.0006, + "loss": 4.091215133666992, + "step": 6778 + }, + { + "epoch": 94.15371179039302, + "grad_norm": 0.016585713252425194, + "learning_rate": 0.0006, + "loss": 4.013662338256836, + "step": 6779 + }, + { + "epoch": 94.16768558951965, + "grad_norm": 0.017397213727235794, + "learning_rate": 0.0006, + "loss": 4.070602893829346, + "step": 6780 + }, + { + "epoch": 94.18165938864628, + "grad_norm": 0.01733497530221939, + "learning_rate": 0.0006, + "loss": 4.1304426193237305, + "step": 6781 + }, + { + "epoch": 94.19563318777293, + "grad_norm": 0.015215300023555756, + "learning_rate": 0.0006, + "loss": 3.965014696121216, + "step": 6782 + }, + { + "epoch": 94.20960698689956, + "grad_norm": 0.015908699482679367, + "learning_rate": 0.0006, + "loss": 3.924964427947998, + "step": 6783 + }, + { + "epoch": 94.2235807860262, + "grad_norm": 0.014044968411326408, + "learning_rate": 0.0006, + "loss": 4.08673095703125, + "step": 6784 + }, + { + "epoch": 94.23755458515284, + "grad_norm": 0.01375142764300108, + "learning_rate": 0.0006, + "loss": 4.011466979980469, + "step": 6785 + }, + { + "epoch": 94.25152838427948, + "grad_norm": 0.015539257787168026, + "learning_rate": 0.0006, + "loss": 4.09976863861084, + "step": 6786 + }, + { + "epoch": 94.26550218340611, + "grad_norm": 0.014303367584943771, + "learning_rate": 0.0006, + "loss": 4.108369827270508, + "step": 6787 + }, + { + "epoch": 94.27947598253274, + "grad_norm": 0.01281928364187479, + "learning_rate": 0.0006, + "loss": 4.073737144470215, + "step": 6788 + }, + { + "epoch": 94.29344978165939, + "grad_norm": 0.013859959319233894, + "learning_rate": 0.0006, + "loss": 4.054823875427246, + "step": 6789 + }, + { + "epoch": 94.30742358078602, + "grad_norm": 0.013326860964298248, + "learning_rate": 0.0006, + "loss": 4.133251190185547, + "step": 6790 + }, + { + "epoch": 94.32139737991267, + "grad_norm": 0.013923076912760735, + "learning_rate": 0.0006, + "loss": 4.190706729888916, + "step": 6791 + }, + { + "epoch": 94.3353711790393, + "grad_norm": 0.01420847699046135, + "learning_rate": 0.0006, + "loss": 3.9336605072021484, + "step": 6792 + }, + { + "epoch": 94.34934497816595, + "grad_norm": 0.014607555232942104, + "learning_rate": 0.0006, + "loss": 4.128399848937988, + "step": 6793 + }, + { + "epoch": 94.36331877729258, + "grad_norm": 0.012862961739301682, + "learning_rate": 0.0006, + "loss": 4.0132527351379395, + "step": 6794 + }, + { + "epoch": 94.37729257641921, + "grad_norm": 0.012530217878520489, + "learning_rate": 0.0006, + "loss": 4.0384368896484375, + "step": 6795 + }, + { + "epoch": 94.39126637554585, + "grad_norm": 0.01286802627146244, + "learning_rate": 0.0006, + "loss": 3.95570707321167, + "step": 6796 + }, + { + "epoch": 94.40524017467249, + "grad_norm": 0.013450917787849903, + "learning_rate": 0.0006, + "loss": 4.016484260559082, + "step": 6797 + }, + { + "epoch": 94.41921397379913, + "grad_norm": 0.01393423043191433, + "learning_rate": 0.0006, + "loss": 4.124959468841553, + "step": 6798 + }, + { + "epoch": 94.43318777292576, + "grad_norm": 0.013077184557914734, + "learning_rate": 0.0006, + "loss": 4.029163837432861, + "step": 6799 + }, + { + "epoch": 94.44716157205241, + "grad_norm": 0.013577334582805634, + "learning_rate": 0.0006, + "loss": 4.22929573059082, + "step": 6800 + }, + { + "epoch": 94.46113537117904, + "grad_norm": 0.013930312357842922, + "learning_rate": 0.0006, + "loss": 4.024711608886719, + "step": 6801 + }, + { + "epoch": 94.47510917030567, + "grad_norm": 0.013746250420808792, + "learning_rate": 0.0006, + "loss": 3.9766428470611572, + "step": 6802 + }, + { + "epoch": 94.48908296943232, + "grad_norm": 0.013682983815670013, + "learning_rate": 0.0006, + "loss": 4.072981834411621, + "step": 6803 + }, + { + "epoch": 94.50305676855895, + "grad_norm": 0.01473842654377222, + "learning_rate": 0.0006, + "loss": 4.197623252868652, + "step": 6804 + }, + { + "epoch": 94.5170305676856, + "grad_norm": 0.014133758842945099, + "learning_rate": 0.0006, + "loss": 4.055549621582031, + "step": 6805 + }, + { + "epoch": 94.53100436681223, + "grad_norm": 0.0139645179733634, + "learning_rate": 0.0006, + "loss": 4.130135536193848, + "step": 6806 + }, + { + "epoch": 94.54497816593886, + "grad_norm": 0.013973365537822247, + "learning_rate": 0.0006, + "loss": 4.006230354309082, + "step": 6807 + }, + { + "epoch": 94.5589519650655, + "grad_norm": 0.014227126725018024, + "learning_rate": 0.0006, + "loss": 4.130459785461426, + "step": 6808 + }, + { + "epoch": 94.57292576419214, + "grad_norm": 0.013129506260156631, + "learning_rate": 0.0006, + "loss": 4.014974594116211, + "step": 6809 + }, + { + "epoch": 94.58689956331878, + "grad_norm": 0.013015891425311565, + "learning_rate": 0.0006, + "loss": 4.161370277404785, + "step": 6810 + }, + { + "epoch": 94.60087336244541, + "grad_norm": 0.013957147486507893, + "learning_rate": 0.0006, + "loss": 4.09519100189209, + "step": 6811 + }, + { + "epoch": 94.61484716157206, + "grad_norm": 0.015494572930037975, + "learning_rate": 0.0006, + "loss": 4.029949188232422, + "step": 6812 + }, + { + "epoch": 94.62882096069869, + "grad_norm": 0.016732869669795036, + "learning_rate": 0.0006, + "loss": 4.135669708251953, + "step": 6813 + }, + { + "epoch": 94.64279475982532, + "grad_norm": 0.01769322343170643, + "learning_rate": 0.0006, + "loss": 4.109872817993164, + "step": 6814 + }, + { + "epoch": 94.65676855895197, + "grad_norm": 0.015932537615299225, + "learning_rate": 0.0006, + "loss": 4.014595031738281, + "step": 6815 + }, + { + "epoch": 94.6707423580786, + "grad_norm": 0.015152990818023682, + "learning_rate": 0.0006, + "loss": 4.1983442306518555, + "step": 6816 + }, + { + "epoch": 94.68471615720524, + "grad_norm": 0.01590685173869133, + "learning_rate": 0.0006, + "loss": 4.119250774383545, + "step": 6817 + }, + { + "epoch": 94.69868995633188, + "grad_norm": 0.016018465161323547, + "learning_rate": 0.0006, + "loss": 4.048069953918457, + "step": 6818 + }, + { + "epoch": 94.71266375545852, + "grad_norm": 0.016621652990579605, + "learning_rate": 0.0006, + "loss": 4.043731689453125, + "step": 6819 + }, + { + "epoch": 94.72663755458515, + "grad_norm": 0.014535024762153625, + "learning_rate": 0.0006, + "loss": 4.1886820793151855, + "step": 6820 + }, + { + "epoch": 94.74061135371178, + "grad_norm": 0.014192107133567333, + "learning_rate": 0.0006, + "loss": 4.103938102722168, + "step": 6821 + }, + { + "epoch": 94.75458515283843, + "grad_norm": 0.0166022852063179, + "learning_rate": 0.0006, + "loss": 4.017818450927734, + "step": 6822 + }, + { + "epoch": 94.76855895196506, + "grad_norm": 0.016247155144810677, + "learning_rate": 0.0006, + "loss": 4.0038981437683105, + "step": 6823 + }, + { + "epoch": 94.78253275109171, + "grad_norm": 0.01593470573425293, + "learning_rate": 0.0006, + "loss": 4.043969631195068, + "step": 6824 + }, + { + "epoch": 94.79650655021834, + "grad_norm": 0.014334345236420631, + "learning_rate": 0.0006, + "loss": 4.097829341888428, + "step": 6825 + }, + { + "epoch": 94.81048034934499, + "grad_norm": 0.015581986866891384, + "learning_rate": 0.0006, + "loss": 4.148984432220459, + "step": 6826 + }, + { + "epoch": 94.82445414847162, + "grad_norm": 0.017519738525152206, + "learning_rate": 0.0006, + "loss": 4.245623588562012, + "step": 6827 + }, + { + "epoch": 94.83842794759825, + "grad_norm": 0.015155277214944363, + "learning_rate": 0.0006, + "loss": 4.031019687652588, + "step": 6828 + }, + { + "epoch": 94.8524017467249, + "grad_norm": 0.014969222247600555, + "learning_rate": 0.0006, + "loss": 4.036084175109863, + "step": 6829 + }, + { + "epoch": 94.86637554585153, + "grad_norm": 0.015987908467650414, + "learning_rate": 0.0006, + "loss": 4.08396053314209, + "step": 6830 + }, + { + "epoch": 94.88034934497817, + "grad_norm": 0.016324151307344437, + "learning_rate": 0.0006, + "loss": 4.036787033081055, + "step": 6831 + }, + { + "epoch": 94.8943231441048, + "grad_norm": 0.015957487747073174, + "learning_rate": 0.0006, + "loss": 4.147607803344727, + "step": 6832 + }, + { + "epoch": 94.90829694323143, + "grad_norm": 0.015286967158317566, + "learning_rate": 0.0006, + "loss": 4.184615135192871, + "step": 6833 + }, + { + "epoch": 94.92227074235808, + "grad_norm": 0.01338377594947815, + "learning_rate": 0.0006, + "loss": 4.014981269836426, + "step": 6834 + }, + { + "epoch": 94.93624454148471, + "grad_norm": 0.013997633941471577, + "learning_rate": 0.0006, + "loss": 4.021933078765869, + "step": 6835 + }, + { + "epoch": 94.95021834061136, + "grad_norm": 0.014572402462363243, + "learning_rate": 0.0006, + "loss": 4.0326948165893555, + "step": 6836 + }, + { + "epoch": 94.96419213973799, + "grad_norm": 0.014379586093127728, + "learning_rate": 0.0006, + "loss": 4.060697555541992, + "step": 6837 + }, + { + "epoch": 94.97816593886463, + "grad_norm": 0.013513042591512203, + "learning_rate": 0.0006, + "loss": 4.05123233795166, + "step": 6838 + }, + { + "epoch": 94.99213973799127, + "grad_norm": 0.012904108501970768, + "learning_rate": 0.0006, + "loss": 4.020852565765381, + "step": 6839 + }, + { + "epoch": 95.0, + "grad_norm": 0.014364209026098251, + "learning_rate": 0.0006, + "loss": 4.080837249755859, + "step": 6840 + }, + { + "epoch": 95.0, + "eval_loss": 4.608776569366455, + "eval_runtime": 56.6516, + "eval_samples_per_second": 43.106, + "eval_steps_per_second": 1.359, + "step": 6840 + }, + { + "epoch": 95.01397379912663, + "grad_norm": 0.01357510220259428, + "learning_rate": 0.0006, + "loss": 4.105312347412109, + "step": 6841 + }, + { + "epoch": 95.02794759825328, + "grad_norm": 0.014248257502913475, + "learning_rate": 0.0006, + "loss": 3.971667766571045, + "step": 6842 + }, + { + "epoch": 95.04192139737991, + "grad_norm": 0.014773483388125896, + "learning_rate": 0.0006, + "loss": 4.03553581237793, + "step": 6843 + }, + { + "epoch": 95.05589519650655, + "grad_norm": 0.014766111969947815, + "learning_rate": 0.0006, + "loss": 4.074707508087158, + "step": 6844 + }, + { + "epoch": 95.06986899563319, + "grad_norm": 0.014778018929064274, + "learning_rate": 0.0006, + "loss": 4.1083855628967285, + "step": 6845 + }, + { + "epoch": 95.08384279475983, + "grad_norm": 0.014466686174273491, + "learning_rate": 0.0006, + "loss": 4.054587364196777, + "step": 6846 + }, + { + "epoch": 95.09781659388646, + "grad_norm": 0.015195904299616814, + "learning_rate": 0.0006, + "loss": 4.142580032348633, + "step": 6847 + }, + { + "epoch": 95.1117903930131, + "grad_norm": 0.01530557032674551, + "learning_rate": 0.0006, + "loss": 4.062134265899658, + "step": 6848 + }, + { + "epoch": 95.12576419213974, + "grad_norm": 0.015203768387436867, + "learning_rate": 0.0006, + "loss": 4.082962989807129, + "step": 6849 + }, + { + "epoch": 95.13973799126637, + "grad_norm": 0.014906628057360649, + "learning_rate": 0.0006, + "loss": 4.033127784729004, + "step": 6850 + }, + { + "epoch": 95.15371179039302, + "grad_norm": 0.01553534809499979, + "learning_rate": 0.0006, + "loss": 4.051743030548096, + "step": 6851 + }, + { + "epoch": 95.16768558951965, + "grad_norm": 0.01607976108789444, + "learning_rate": 0.0006, + "loss": 4.038073539733887, + "step": 6852 + }, + { + "epoch": 95.18165938864628, + "grad_norm": 0.015776360407471657, + "learning_rate": 0.0006, + "loss": 4.013091087341309, + "step": 6853 + }, + { + "epoch": 95.19563318777293, + "grad_norm": 0.01623626798391342, + "learning_rate": 0.0006, + "loss": 4.12293815612793, + "step": 6854 + }, + { + "epoch": 95.20960698689956, + "grad_norm": 0.016887828707695007, + "learning_rate": 0.0006, + "loss": 4.120452404022217, + "step": 6855 + }, + { + "epoch": 95.2235807860262, + "grad_norm": 0.01561128068715334, + "learning_rate": 0.0006, + "loss": 4.071322441101074, + "step": 6856 + }, + { + "epoch": 95.23755458515284, + "grad_norm": 0.015679724514484406, + "learning_rate": 0.0006, + "loss": 4.065072059631348, + "step": 6857 + }, + { + "epoch": 95.25152838427948, + "grad_norm": 0.016506962478160858, + "learning_rate": 0.0006, + "loss": 4.0667924880981445, + "step": 6858 + }, + { + "epoch": 95.26550218340611, + "grad_norm": 0.015340839512646198, + "learning_rate": 0.0006, + "loss": 4.14047908782959, + "step": 6859 + }, + { + "epoch": 95.27947598253274, + "grad_norm": 0.013935413211584091, + "learning_rate": 0.0006, + "loss": 3.9688191413879395, + "step": 6860 + }, + { + "epoch": 95.29344978165939, + "grad_norm": 0.015223405323922634, + "learning_rate": 0.0006, + "loss": 4.101613521575928, + "step": 6861 + }, + { + "epoch": 95.30742358078602, + "grad_norm": 0.01562546379864216, + "learning_rate": 0.0006, + "loss": 4.128626823425293, + "step": 6862 + }, + { + "epoch": 95.32139737991267, + "grad_norm": 0.016159312799572945, + "learning_rate": 0.0006, + "loss": 4.080667018890381, + "step": 6863 + }, + { + "epoch": 95.3353711790393, + "grad_norm": 0.01724511943757534, + "learning_rate": 0.0006, + "loss": 4.08463191986084, + "step": 6864 + }, + { + "epoch": 95.34934497816595, + "grad_norm": 0.01592276245355606, + "learning_rate": 0.0006, + "loss": 4.127612113952637, + "step": 6865 + }, + { + "epoch": 95.36331877729258, + "grad_norm": 0.01744082197546959, + "learning_rate": 0.0006, + "loss": 4.034480094909668, + "step": 6866 + }, + { + "epoch": 95.37729257641921, + "grad_norm": 0.01774069108068943, + "learning_rate": 0.0006, + "loss": 4.091196537017822, + "step": 6867 + }, + { + "epoch": 95.39126637554585, + "grad_norm": 0.01653675176203251, + "learning_rate": 0.0006, + "loss": 4.032937526702881, + "step": 6868 + }, + { + "epoch": 95.40524017467249, + "grad_norm": 0.014933439902961254, + "learning_rate": 0.0006, + "loss": 3.9386508464813232, + "step": 6869 + }, + { + "epoch": 95.41921397379913, + "grad_norm": 0.014743257313966751, + "learning_rate": 0.0006, + "loss": 4.126462459564209, + "step": 6870 + }, + { + "epoch": 95.43318777292576, + "grad_norm": 0.014614643529057503, + "learning_rate": 0.0006, + "loss": 4.093838214874268, + "step": 6871 + }, + { + "epoch": 95.44716157205241, + "grad_norm": 0.01665084809064865, + "learning_rate": 0.0006, + "loss": 4.1471967697143555, + "step": 6872 + }, + { + "epoch": 95.46113537117904, + "grad_norm": 0.018056001514196396, + "learning_rate": 0.0006, + "loss": 4.1388349533081055, + "step": 6873 + }, + { + "epoch": 95.47510917030567, + "grad_norm": 0.02129282057285309, + "learning_rate": 0.0006, + "loss": 4.130726337432861, + "step": 6874 + }, + { + "epoch": 95.48908296943232, + "grad_norm": 0.022237276658415794, + "learning_rate": 0.0006, + "loss": 4.10750675201416, + "step": 6875 + }, + { + "epoch": 95.50305676855895, + "grad_norm": 0.021106960251927376, + "learning_rate": 0.0006, + "loss": 4.136451721191406, + "step": 6876 + }, + { + "epoch": 95.5170305676856, + "grad_norm": 0.019107872620224953, + "learning_rate": 0.0006, + "loss": 4.139644145965576, + "step": 6877 + }, + { + "epoch": 95.53100436681223, + "grad_norm": 0.018290145322680473, + "learning_rate": 0.0006, + "loss": 4.161884784698486, + "step": 6878 + }, + { + "epoch": 95.54497816593886, + "grad_norm": 0.018161306157708168, + "learning_rate": 0.0006, + "loss": 4.1252336502075195, + "step": 6879 + }, + { + "epoch": 95.5589519650655, + "grad_norm": 0.018432756885886192, + "learning_rate": 0.0006, + "loss": 4.0635986328125, + "step": 6880 + }, + { + "epoch": 95.57292576419214, + "grad_norm": 0.016592558473348618, + "learning_rate": 0.0006, + "loss": 4.037112236022949, + "step": 6881 + }, + { + "epoch": 95.58689956331878, + "grad_norm": 0.015283452346920967, + "learning_rate": 0.0006, + "loss": 4.1964826583862305, + "step": 6882 + }, + { + "epoch": 95.60087336244541, + "grad_norm": 0.015539180487394333, + "learning_rate": 0.0006, + "loss": 4.110917568206787, + "step": 6883 + }, + { + "epoch": 95.61484716157206, + "grad_norm": 0.016508113592863083, + "learning_rate": 0.0006, + "loss": 3.9705517292022705, + "step": 6884 + }, + { + "epoch": 95.62882096069869, + "grad_norm": 0.016635406762361526, + "learning_rate": 0.0006, + "loss": 4.025636196136475, + "step": 6885 + }, + { + "epoch": 95.64279475982532, + "grad_norm": 0.01688900962471962, + "learning_rate": 0.0006, + "loss": 4.01814079284668, + "step": 6886 + }, + { + "epoch": 95.65676855895197, + "grad_norm": 0.01614062674343586, + "learning_rate": 0.0006, + "loss": 4.098041534423828, + "step": 6887 + }, + { + "epoch": 95.6707423580786, + "grad_norm": 0.015679053962230682, + "learning_rate": 0.0006, + "loss": 4.077049255371094, + "step": 6888 + }, + { + "epoch": 95.68471615720524, + "grad_norm": 0.016617605462670326, + "learning_rate": 0.0006, + "loss": 4.10363245010376, + "step": 6889 + }, + { + "epoch": 95.69868995633188, + "grad_norm": 0.015966709703207016, + "learning_rate": 0.0006, + "loss": 4.100924491882324, + "step": 6890 + }, + { + "epoch": 95.71266375545852, + "grad_norm": 0.015768567100167274, + "learning_rate": 0.0006, + "loss": 4.063976287841797, + "step": 6891 + }, + { + "epoch": 95.72663755458515, + "grad_norm": 0.015800341963768005, + "learning_rate": 0.0006, + "loss": 4.008409023284912, + "step": 6892 + }, + { + "epoch": 95.74061135371178, + "grad_norm": 0.014288571663200855, + "learning_rate": 0.0006, + "loss": 4.139591217041016, + "step": 6893 + }, + { + "epoch": 95.75458515283843, + "grad_norm": 0.013743448071181774, + "learning_rate": 0.0006, + "loss": 3.967352867126465, + "step": 6894 + }, + { + "epoch": 95.76855895196506, + "grad_norm": 0.01387989055365324, + "learning_rate": 0.0006, + "loss": 4.127554893493652, + "step": 6895 + }, + { + "epoch": 95.78253275109171, + "grad_norm": 0.013910802081227303, + "learning_rate": 0.0006, + "loss": 4.130941390991211, + "step": 6896 + }, + { + "epoch": 95.79650655021834, + "grad_norm": 0.013936805538833141, + "learning_rate": 0.0006, + "loss": 4.138744831085205, + "step": 6897 + }, + { + "epoch": 95.81048034934499, + "grad_norm": 0.015135723166167736, + "learning_rate": 0.0006, + "loss": 4.131860256195068, + "step": 6898 + }, + { + "epoch": 95.82445414847162, + "grad_norm": 0.014921358786523342, + "learning_rate": 0.0006, + "loss": 4.1949462890625, + "step": 6899 + }, + { + "epoch": 95.83842794759825, + "grad_norm": 0.01401488296687603, + "learning_rate": 0.0006, + "loss": 4.181267738342285, + "step": 6900 + }, + { + "epoch": 95.8524017467249, + "grad_norm": 0.013691626489162445, + "learning_rate": 0.0006, + "loss": 4.113667964935303, + "step": 6901 + }, + { + "epoch": 95.86637554585153, + "grad_norm": 0.013890047557651997, + "learning_rate": 0.0006, + "loss": 4.163067817687988, + "step": 6902 + }, + { + "epoch": 95.88034934497817, + "grad_norm": 0.01385503076016903, + "learning_rate": 0.0006, + "loss": 4.076153755187988, + "step": 6903 + }, + { + "epoch": 95.8943231441048, + "grad_norm": 0.01383926346898079, + "learning_rate": 0.0006, + "loss": 4.102423667907715, + "step": 6904 + }, + { + "epoch": 95.90829694323143, + "grad_norm": 0.015961717814207077, + "learning_rate": 0.0006, + "loss": 4.1543779373168945, + "step": 6905 + }, + { + "epoch": 95.92227074235808, + "grad_norm": 0.016718363389372826, + "learning_rate": 0.0006, + "loss": 3.991607189178467, + "step": 6906 + }, + { + "epoch": 95.93624454148471, + "grad_norm": 0.01585792563855648, + "learning_rate": 0.0006, + "loss": 4.048587799072266, + "step": 6907 + }, + { + "epoch": 95.95021834061136, + "grad_norm": 0.015311713330447674, + "learning_rate": 0.0006, + "loss": 4.1070451736450195, + "step": 6908 + }, + { + "epoch": 95.96419213973799, + "grad_norm": 0.014521263539791107, + "learning_rate": 0.0006, + "loss": 4.087587356567383, + "step": 6909 + }, + { + "epoch": 95.97816593886463, + "grad_norm": 0.01433157455176115, + "learning_rate": 0.0006, + "loss": 4.139904499053955, + "step": 6910 + }, + { + "epoch": 95.99213973799127, + "grad_norm": 0.014475514180958271, + "learning_rate": 0.0006, + "loss": 4.182525634765625, + "step": 6911 + }, + { + "epoch": 96.0, + "grad_norm": 0.015245620161294937, + "learning_rate": 0.0006, + "loss": 4.1199870109558105, + "step": 6912 + }, + { + "epoch": 96.0, + "eval_loss": 4.58721399307251, + "eval_runtime": 56.395, + "eval_samples_per_second": 43.302, + "eval_steps_per_second": 1.365, + "step": 6912 + }, + { + "epoch": 96.01397379912663, + "grad_norm": 0.014325885102152824, + "learning_rate": 0.0006, + "loss": 4.025214195251465, + "step": 6913 + }, + { + "epoch": 96.02794759825328, + "grad_norm": 0.01678297482430935, + "learning_rate": 0.0006, + "loss": 3.998842716217041, + "step": 6914 + }, + { + "epoch": 96.04192139737991, + "grad_norm": 0.017044762149453163, + "learning_rate": 0.0006, + "loss": 4.122270584106445, + "step": 6915 + }, + { + "epoch": 96.05589519650655, + "grad_norm": 0.017583075910806656, + "learning_rate": 0.0006, + "loss": 3.9782891273498535, + "step": 6916 + }, + { + "epoch": 96.06986899563319, + "grad_norm": 0.0179787315428257, + "learning_rate": 0.0006, + "loss": 4.030402660369873, + "step": 6917 + }, + { + "epoch": 96.08384279475983, + "grad_norm": 0.0159553661942482, + "learning_rate": 0.0006, + "loss": 3.949040651321411, + "step": 6918 + }, + { + "epoch": 96.09781659388646, + "grad_norm": 0.01554026547819376, + "learning_rate": 0.0006, + "loss": 4.0429534912109375, + "step": 6919 + }, + { + "epoch": 96.1117903930131, + "grad_norm": 0.016099359840154648, + "learning_rate": 0.0006, + "loss": 4.037537097930908, + "step": 6920 + }, + { + "epoch": 96.12576419213974, + "grad_norm": 0.01676846854388714, + "learning_rate": 0.0006, + "loss": 4.022023677825928, + "step": 6921 + }, + { + "epoch": 96.13973799126637, + "grad_norm": 0.019358212128281593, + "learning_rate": 0.0006, + "loss": 4.048679351806641, + "step": 6922 + }, + { + "epoch": 96.15371179039302, + "grad_norm": 0.020390303805470467, + "learning_rate": 0.0006, + "loss": 4.144132614135742, + "step": 6923 + }, + { + "epoch": 96.16768558951965, + "grad_norm": 0.01884407177567482, + "learning_rate": 0.0006, + "loss": 4.128715515136719, + "step": 6924 + }, + { + "epoch": 96.18165938864628, + "grad_norm": 0.015098451636731625, + "learning_rate": 0.0006, + "loss": 4.128741264343262, + "step": 6925 + }, + { + "epoch": 96.19563318777293, + "grad_norm": 0.0148141635581851, + "learning_rate": 0.0006, + "loss": 4.054587364196777, + "step": 6926 + }, + { + "epoch": 96.20960698689956, + "grad_norm": 0.015594304539263248, + "learning_rate": 0.0006, + "loss": 4.073890209197998, + "step": 6927 + }, + { + "epoch": 96.2235807860262, + "grad_norm": 0.0139383003115654, + "learning_rate": 0.0006, + "loss": 4.115614891052246, + "step": 6928 + }, + { + "epoch": 96.23755458515284, + "grad_norm": 0.0150460135191679, + "learning_rate": 0.0006, + "loss": 4.040738105773926, + "step": 6929 + }, + { + "epoch": 96.25152838427948, + "grad_norm": 0.013879122212529182, + "learning_rate": 0.0006, + "loss": 3.9292469024658203, + "step": 6930 + }, + { + "epoch": 96.26550218340611, + "grad_norm": 0.015546537935733795, + "learning_rate": 0.0006, + "loss": 4.111361503601074, + "step": 6931 + }, + { + "epoch": 96.27947598253274, + "grad_norm": 0.016172925010323524, + "learning_rate": 0.0006, + "loss": 4.041029930114746, + "step": 6932 + }, + { + "epoch": 96.29344978165939, + "grad_norm": 0.015113193541765213, + "learning_rate": 0.0006, + "loss": 4.157950401306152, + "step": 6933 + }, + { + "epoch": 96.30742358078602, + "grad_norm": 0.014663247391581535, + "learning_rate": 0.0006, + "loss": 4.020952224731445, + "step": 6934 + }, + { + "epoch": 96.32139737991267, + "grad_norm": 0.015092858113348484, + "learning_rate": 0.0006, + "loss": 4.065241813659668, + "step": 6935 + }, + { + "epoch": 96.3353711790393, + "grad_norm": 0.013199460692703724, + "learning_rate": 0.0006, + "loss": 4.062674522399902, + "step": 6936 + }, + { + "epoch": 96.34934497816595, + "grad_norm": 0.015065675601363182, + "learning_rate": 0.0006, + "loss": 4.038748741149902, + "step": 6937 + }, + { + "epoch": 96.36331877729258, + "grad_norm": 0.014270043931901455, + "learning_rate": 0.0006, + "loss": 4.073092460632324, + "step": 6938 + }, + { + "epoch": 96.37729257641921, + "grad_norm": 0.01423644833266735, + "learning_rate": 0.0006, + "loss": 4.238361835479736, + "step": 6939 + }, + { + "epoch": 96.39126637554585, + "grad_norm": 0.01380224246531725, + "learning_rate": 0.0006, + "loss": 4.137866973876953, + "step": 6940 + }, + { + "epoch": 96.40524017467249, + "grad_norm": 0.012464135885238647, + "learning_rate": 0.0006, + "loss": 3.9527339935302734, + "step": 6941 + }, + { + "epoch": 96.41921397379913, + "grad_norm": 0.013094968162477016, + "learning_rate": 0.0006, + "loss": 4.106327056884766, + "step": 6942 + }, + { + "epoch": 96.43318777292576, + "grad_norm": 0.013918030075728893, + "learning_rate": 0.0006, + "loss": 4.072292327880859, + "step": 6943 + }, + { + "epoch": 96.44716157205241, + "grad_norm": 0.013835827820003033, + "learning_rate": 0.0006, + "loss": 4.094313621520996, + "step": 6944 + }, + { + "epoch": 96.46113537117904, + "grad_norm": 0.013890027068555355, + "learning_rate": 0.0006, + "loss": 3.947615623474121, + "step": 6945 + }, + { + "epoch": 96.47510917030567, + "grad_norm": 0.012371144257485867, + "learning_rate": 0.0006, + "loss": 4.038928985595703, + "step": 6946 + }, + { + "epoch": 96.48908296943232, + "grad_norm": 0.012236416339874268, + "learning_rate": 0.0006, + "loss": 3.9835925102233887, + "step": 6947 + }, + { + "epoch": 96.50305676855895, + "grad_norm": 0.012866619043052197, + "learning_rate": 0.0006, + "loss": 4.076041221618652, + "step": 6948 + }, + { + "epoch": 96.5170305676856, + "grad_norm": 0.013425248675048351, + "learning_rate": 0.0006, + "loss": 4.086307525634766, + "step": 6949 + }, + { + "epoch": 96.53100436681223, + "grad_norm": 0.014234314672648907, + "learning_rate": 0.0006, + "loss": 4.087897300720215, + "step": 6950 + }, + { + "epoch": 96.54497816593886, + "grad_norm": 0.013670518063008785, + "learning_rate": 0.0006, + "loss": 4.088400840759277, + "step": 6951 + }, + { + "epoch": 96.5589519650655, + "grad_norm": 0.013920796103775501, + "learning_rate": 0.0006, + "loss": 4.026619911193848, + "step": 6952 + }, + { + "epoch": 96.57292576419214, + "grad_norm": 0.013115637935698032, + "learning_rate": 0.0006, + "loss": 4.141090393066406, + "step": 6953 + }, + { + "epoch": 96.58689956331878, + "grad_norm": 0.013466455042362213, + "learning_rate": 0.0006, + "loss": 4.048084259033203, + "step": 6954 + }, + { + "epoch": 96.60087336244541, + "grad_norm": 0.013038012199103832, + "learning_rate": 0.0006, + "loss": 4.034649848937988, + "step": 6955 + }, + { + "epoch": 96.61484716157206, + "grad_norm": 0.012926512397825718, + "learning_rate": 0.0006, + "loss": 4.138445854187012, + "step": 6956 + }, + { + "epoch": 96.62882096069869, + "grad_norm": 0.013271111994981766, + "learning_rate": 0.0006, + "loss": 4.240598201751709, + "step": 6957 + }, + { + "epoch": 96.64279475982532, + "grad_norm": 0.013092855922877789, + "learning_rate": 0.0006, + "loss": 4.091870307922363, + "step": 6958 + }, + { + "epoch": 96.65676855895197, + "grad_norm": 0.013304130174219608, + "learning_rate": 0.0006, + "loss": 4.048708915710449, + "step": 6959 + }, + { + "epoch": 96.6707423580786, + "grad_norm": 0.01289574895054102, + "learning_rate": 0.0006, + "loss": 4.083846092224121, + "step": 6960 + }, + { + "epoch": 96.68471615720524, + "grad_norm": 0.012587146833539009, + "learning_rate": 0.0006, + "loss": 4.1233015060424805, + "step": 6961 + }, + { + "epoch": 96.69868995633188, + "grad_norm": 0.012312786653637886, + "learning_rate": 0.0006, + "loss": 4.0938262939453125, + "step": 6962 + }, + { + "epoch": 96.71266375545852, + "grad_norm": 0.014632516540586948, + "learning_rate": 0.0006, + "loss": 4.0226545333862305, + "step": 6963 + }, + { + "epoch": 96.72663755458515, + "grad_norm": 0.015110369771718979, + "learning_rate": 0.0006, + "loss": 4.052829742431641, + "step": 6964 + }, + { + "epoch": 96.74061135371178, + "grad_norm": 0.015997016802430153, + "learning_rate": 0.0006, + "loss": 4.126044750213623, + "step": 6965 + }, + { + "epoch": 96.75458515283843, + "grad_norm": 0.016428956761956215, + "learning_rate": 0.0006, + "loss": 4.126011848449707, + "step": 6966 + }, + { + "epoch": 96.76855895196506, + "grad_norm": 0.01638922467827797, + "learning_rate": 0.0006, + "loss": 4.1207685470581055, + "step": 6967 + }, + { + "epoch": 96.78253275109171, + "grad_norm": 0.01387663371860981, + "learning_rate": 0.0006, + "loss": 3.9403412342071533, + "step": 6968 + }, + { + "epoch": 96.79650655021834, + "grad_norm": 0.014120660722255707, + "learning_rate": 0.0006, + "loss": 4.202868461608887, + "step": 6969 + }, + { + "epoch": 96.81048034934499, + "grad_norm": 0.014854361303150654, + "learning_rate": 0.0006, + "loss": 4.015185832977295, + "step": 6970 + }, + { + "epoch": 96.82445414847162, + "grad_norm": 0.014359727501869202, + "learning_rate": 0.0006, + "loss": 4.119946479797363, + "step": 6971 + }, + { + "epoch": 96.83842794759825, + "grad_norm": 0.01564539223909378, + "learning_rate": 0.0006, + "loss": 3.9208037853240967, + "step": 6972 + }, + { + "epoch": 96.8524017467249, + "grad_norm": 0.015814676880836487, + "learning_rate": 0.0006, + "loss": 4.056029319763184, + "step": 6973 + }, + { + "epoch": 96.86637554585153, + "grad_norm": 0.016342822462320328, + "learning_rate": 0.0006, + "loss": 4.160576343536377, + "step": 6974 + }, + { + "epoch": 96.88034934497817, + "grad_norm": 0.016163306310772896, + "learning_rate": 0.0006, + "loss": 3.94950008392334, + "step": 6975 + }, + { + "epoch": 96.8943231441048, + "grad_norm": 0.018308300524950027, + "learning_rate": 0.0006, + "loss": 4.096719741821289, + "step": 6976 + }, + { + "epoch": 96.90829694323143, + "grad_norm": 0.01985356956720352, + "learning_rate": 0.0006, + "loss": 4.052276611328125, + "step": 6977 + }, + { + "epoch": 96.92227074235808, + "grad_norm": 0.01949184387922287, + "learning_rate": 0.0006, + "loss": 4.077932357788086, + "step": 6978 + }, + { + "epoch": 96.93624454148471, + "grad_norm": 0.018964678049087524, + "learning_rate": 0.0006, + "loss": 4.019469738006592, + "step": 6979 + }, + { + "epoch": 96.95021834061136, + "grad_norm": 0.019365238025784492, + "learning_rate": 0.0006, + "loss": 4.032814979553223, + "step": 6980 + }, + { + "epoch": 96.96419213973799, + "grad_norm": 0.01766037568449974, + "learning_rate": 0.0006, + "loss": 4.061336517333984, + "step": 6981 + }, + { + "epoch": 96.97816593886463, + "grad_norm": 0.018002403900027275, + "learning_rate": 0.0006, + "loss": 4.054131507873535, + "step": 6982 + }, + { + "epoch": 96.99213973799127, + "grad_norm": 0.01888604648411274, + "learning_rate": 0.0006, + "loss": 3.9295921325683594, + "step": 6983 + }, + { + "epoch": 97.0, + "grad_norm": 0.01609404757618904, + "learning_rate": 0.0006, + "loss": 4.037106513977051, + "step": 6984 + }, + { + "epoch": 97.0, + "eval_loss": 4.5986433029174805, + "eval_runtime": 56.5822, + "eval_samples_per_second": 43.158, + "eval_steps_per_second": 1.361, + "step": 6984 + }, + { + "epoch": 97.01397379912663, + "grad_norm": 0.015459981746971607, + "learning_rate": 0.0006, + "loss": 4.06088924407959, + "step": 6985 + }, + { + "epoch": 97.02794759825328, + "grad_norm": 0.016333693638443947, + "learning_rate": 0.0006, + "loss": 3.9525675773620605, + "step": 6986 + }, + { + "epoch": 97.04192139737991, + "grad_norm": 0.0157504603266716, + "learning_rate": 0.0006, + "loss": 4.143257141113281, + "step": 6987 + }, + { + "epoch": 97.05589519650655, + "grad_norm": 0.015252048149704933, + "learning_rate": 0.0006, + "loss": 4.158165454864502, + "step": 6988 + }, + { + "epoch": 97.06986899563319, + "grad_norm": 0.015041683800518513, + "learning_rate": 0.0006, + "loss": 4.030884742736816, + "step": 6989 + }, + { + "epoch": 97.08384279475983, + "grad_norm": 0.014693699777126312, + "learning_rate": 0.0006, + "loss": 3.8566339015960693, + "step": 6990 + }, + { + "epoch": 97.09781659388646, + "grad_norm": 0.0141681469976902, + "learning_rate": 0.0006, + "loss": 4.001909255981445, + "step": 6991 + }, + { + "epoch": 97.1117903930131, + "grad_norm": 0.013231917284429073, + "learning_rate": 0.0006, + "loss": 3.995041608810425, + "step": 6992 + }, + { + "epoch": 97.12576419213974, + "grad_norm": 0.014617119915783405, + "learning_rate": 0.0006, + "loss": 4.012053489685059, + "step": 6993 + }, + { + "epoch": 97.13973799126637, + "grad_norm": 0.014983531087636948, + "learning_rate": 0.0006, + "loss": 4.084125518798828, + "step": 6994 + }, + { + "epoch": 97.15371179039302, + "grad_norm": 0.013970102183520794, + "learning_rate": 0.0006, + "loss": 4.001402378082275, + "step": 6995 + }, + { + "epoch": 97.16768558951965, + "grad_norm": 0.0134900426492095, + "learning_rate": 0.0006, + "loss": 4.007713794708252, + "step": 6996 + }, + { + "epoch": 97.18165938864628, + "grad_norm": 0.013806418515741825, + "learning_rate": 0.0006, + "loss": 3.9524781703948975, + "step": 6997 + }, + { + "epoch": 97.19563318777293, + "grad_norm": 0.014304264448583126, + "learning_rate": 0.0006, + "loss": 3.9666526317596436, + "step": 6998 + }, + { + "epoch": 97.20960698689956, + "grad_norm": 0.016392510384321213, + "learning_rate": 0.0006, + "loss": 3.9915683269500732, + "step": 6999 + }, + { + "epoch": 97.2235807860262, + "grad_norm": 0.017044387757778168, + "learning_rate": 0.0006, + "loss": 4.11865234375, + "step": 7000 + }, + { + "epoch": 97.23755458515284, + "grad_norm": 0.016325082629919052, + "learning_rate": 0.0006, + "loss": 4.0636820793151855, + "step": 7001 + }, + { + "epoch": 97.25152838427948, + "grad_norm": 0.014634544961154461, + "learning_rate": 0.0006, + "loss": 3.967127799987793, + "step": 7002 + }, + { + "epoch": 97.26550218340611, + "grad_norm": 0.014043385162949562, + "learning_rate": 0.0006, + "loss": 4.072741985321045, + "step": 7003 + }, + { + "epoch": 97.27947598253274, + "grad_norm": 0.01376431342214346, + "learning_rate": 0.0006, + "loss": 4.07547664642334, + "step": 7004 + }, + { + "epoch": 97.29344978165939, + "grad_norm": 0.013959003612399101, + "learning_rate": 0.0006, + "loss": 3.9862523078918457, + "step": 7005 + }, + { + "epoch": 97.30742358078602, + "grad_norm": 0.01420508325099945, + "learning_rate": 0.0006, + "loss": 4.085540294647217, + "step": 7006 + }, + { + "epoch": 97.32139737991267, + "grad_norm": 0.01613876223564148, + "learning_rate": 0.0006, + "loss": 4.025545120239258, + "step": 7007 + }, + { + "epoch": 97.3353711790393, + "grad_norm": 0.017000367864966393, + "learning_rate": 0.0006, + "loss": 4.047660827636719, + "step": 7008 + }, + { + "epoch": 97.34934497816595, + "grad_norm": 0.01646680012345314, + "learning_rate": 0.0006, + "loss": 4.087928771972656, + "step": 7009 + }, + { + "epoch": 97.36331877729258, + "grad_norm": 0.015728263184428215, + "learning_rate": 0.0006, + "loss": 4.022225856781006, + "step": 7010 + }, + { + "epoch": 97.37729257641921, + "grad_norm": 0.015262200497090816, + "learning_rate": 0.0006, + "loss": 3.944697618484497, + "step": 7011 + }, + { + "epoch": 97.39126637554585, + "grad_norm": 0.016452079638838768, + "learning_rate": 0.0006, + "loss": 3.957644462585449, + "step": 7012 + }, + { + "epoch": 97.40524017467249, + "grad_norm": 0.018871447071433067, + "learning_rate": 0.0006, + "loss": 4.017784118652344, + "step": 7013 + }, + { + "epoch": 97.41921397379913, + "grad_norm": 0.01902632787823677, + "learning_rate": 0.0006, + "loss": 3.973320484161377, + "step": 7014 + }, + { + "epoch": 97.43318777292576, + "grad_norm": 0.016632359474897385, + "learning_rate": 0.0006, + "loss": 4.047779083251953, + "step": 7015 + }, + { + "epoch": 97.44716157205241, + "grad_norm": 0.01676122285425663, + "learning_rate": 0.0006, + "loss": 4.149074554443359, + "step": 7016 + }, + { + "epoch": 97.46113537117904, + "grad_norm": 0.01764233037829399, + "learning_rate": 0.0006, + "loss": 3.92502760887146, + "step": 7017 + }, + { + "epoch": 97.47510917030567, + "grad_norm": 0.01836075820028782, + "learning_rate": 0.0006, + "loss": 4.173796653747559, + "step": 7018 + }, + { + "epoch": 97.48908296943232, + "grad_norm": 0.017320366576313972, + "learning_rate": 0.0006, + "loss": 4.007115364074707, + "step": 7019 + }, + { + "epoch": 97.50305676855895, + "grad_norm": 0.015976276248693466, + "learning_rate": 0.0006, + "loss": 4.165112495422363, + "step": 7020 + }, + { + "epoch": 97.5170305676856, + "grad_norm": 0.0173528753221035, + "learning_rate": 0.0006, + "loss": 4.158927917480469, + "step": 7021 + }, + { + "epoch": 97.53100436681223, + "grad_norm": 0.01881110668182373, + "learning_rate": 0.0006, + "loss": 4.069684982299805, + "step": 7022 + }, + { + "epoch": 97.54497816593886, + "grad_norm": 0.01849045790731907, + "learning_rate": 0.0006, + "loss": 4.085758209228516, + "step": 7023 + }, + { + "epoch": 97.5589519650655, + "grad_norm": 0.016648925840854645, + "learning_rate": 0.0006, + "loss": 4.047055721282959, + "step": 7024 + }, + { + "epoch": 97.57292576419214, + "grad_norm": 0.014700878411531448, + "learning_rate": 0.0006, + "loss": 4.0177507400512695, + "step": 7025 + }, + { + "epoch": 97.58689956331878, + "grad_norm": 0.014601817354559898, + "learning_rate": 0.0006, + "loss": 4.070394039154053, + "step": 7026 + }, + { + "epoch": 97.60087336244541, + "grad_norm": 0.016391467303037643, + "learning_rate": 0.0006, + "loss": 3.957401990890503, + "step": 7027 + }, + { + "epoch": 97.61484716157206, + "grad_norm": 0.015401630662381649, + "learning_rate": 0.0006, + "loss": 4.040555000305176, + "step": 7028 + }, + { + "epoch": 97.62882096069869, + "grad_norm": 0.014783227816224098, + "learning_rate": 0.0006, + "loss": 4.039798259735107, + "step": 7029 + }, + { + "epoch": 97.64279475982532, + "grad_norm": 0.014704087749123573, + "learning_rate": 0.0006, + "loss": 4.059834003448486, + "step": 7030 + }, + { + "epoch": 97.65676855895197, + "grad_norm": 0.01375004556030035, + "learning_rate": 0.0006, + "loss": 4.084839344024658, + "step": 7031 + }, + { + "epoch": 97.6707423580786, + "grad_norm": 0.01403744611889124, + "learning_rate": 0.0006, + "loss": 4.060206890106201, + "step": 7032 + }, + { + "epoch": 97.68471615720524, + "grad_norm": 0.012827585451304913, + "learning_rate": 0.0006, + "loss": 4.187256813049316, + "step": 7033 + }, + { + "epoch": 97.69868995633188, + "grad_norm": 0.013072148896753788, + "learning_rate": 0.0006, + "loss": 4.136262893676758, + "step": 7034 + }, + { + "epoch": 97.71266375545852, + "grad_norm": 0.013496107421815395, + "learning_rate": 0.0006, + "loss": 3.9419479370117188, + "step": 7035 + }, + { + "epoch": 97.72663755458515, + "grad_norm": 0.014445106498897076, + "learning_rate": 0.0006, + "loss": 3.9494845867156982, + "step": 7036 + }, + { + "epoch": 97.74061135371178, + "grad_norm": 0.014507179148495197, + "learning_rate": 0.0006, + "loss": 4.05567741394043, + "step": 7037 + }, + { + "epoch": 97.75458515283843, + "grad_norm": 0.015422643162310123, + "learning_rate": 0.0006, + "loss": 4.103999137878418, + "step": 7038 + }, + { + "epoch": 97.76855895196506, + "grad_norm": 0.01515049859881401, + "learning_rate": 0.0006, + "loss": 4.094231605529785, + "step": 7039 + }, + { + "epoch": 97.78253275109171, + "grad_norm": 0.013850145973265171, + "learning_rate": 0.0006, + "loss": 4.084870338439941, + "step": 7040 + }, + { + "epoch": 97.79650655021834, + "grad_norm": 0.013332060538232327, + "learning_rate": 0.0006, + "loss": 4.193096160888672, + "step": 7041 + }, + { + "epoch": 97.81048034934499, + "grad_norm": 0.014778176322579384, + "learning_rate": 0.0006, + "loss": 4.0440216064453125, + "step": 7042 + }, + { + "epoch": 97.82445414847162, + "grad_norm": 0.013498589396476746, + "learning_rate": 0.0006, + "loss": 4.088800430297852, + "step": 7043 + }, + { + "epoch": 97.83842794759825, + "grad_norm": 0.014520619064569473, + "learning_rate": 0.0006, + "loss": 4.156583786010742, + "step": 7044 + }, + { + "epoch": 97.8524017467249, + "grad_norm": 0.014858689159154892, + "learning_rate": 0.0006, + "loss": 4.113174915313721, + "step": 7045 + }, + { + "epoch": 97.86637554585153, + "grad_norm": 0.015516536310315132, + "learning_rate": 0.0006, + "loss": 4.07958984375, + "step": 7046 + }, + { + "epoch": 97.88034934497817, + "grad_norm": 0.016619030386209488, + "learning_rate": 0.0006, + "loss": 4.03786563873291, + "step": 7047 + }, + { + "epoch": 97.8943231441048, + "grad_norm": 0.015789631754159927, + "learning_rate": 0.0006, + "loss": 4.065053462982178, + "step": 7048 + }, + { + "epoch": 97.90829694323143, + "grad_norm": 0.015848442912101746, + "learning_rate": 0.0006, + "loss": 4.183546543121338, + "step": 7049 + }, + { + "epoch": 97.92227074235808, + "grad_norm": 0.014621895737946033, + "learning_rate": 0.0006, + "loss": 4.024744987487793, + "step": 7050 + }, + { + "epoch": 97.93624454148471, + "grad_norm": 0.015793966129422188, + "learning_rate": 0.0006, + "loss": 3.9715819358825684, + "step": 7051 + }, + { + "epoch": 97.95021834061136, + "grad_norm": 0.014637742191553116, + "learning_rate": 0.0006, + "loss": 4.061223983764648, + "step": 7052 + }, + { + "epoch": 97.96419213973799, + "grad_norm": 0.0143470773473382, + "learning_rate": 0.0006, + "loss": 4.00929069519043, + "step": 7053 + }, + { + "epoch": 97.97816593886463, + "grad_norm": 0.014167622663080692, + "learning_rate": 0.0006, + "loss": 3.982048749923706, + "step": 7054 + }, + { + "epoch": 97.99213973799127, + "grad_norm": 0.01527823694050312, + "learning_rate": 0.0006, + "loss": 4.063419818878174, + "step": 7055 + }, + { + "epoch": 98.0, + "grad_norm": 0.015410988591611385, + "learning_rate": 0.0006, + "loss": 4.058195114135742, + "step": 7056 + }, + { + "epoch": 98.0, + "eval_loss": 4.5309038162231445, + "eval_runtime": 56.6931, + "eval_samples_per_second": 43.074, + "eval_steps_per_second": 1.358, + "step": 7056 + }, + { + "epoch": 98.01397379912663, + "grad_norm": 0.013760345987975597, + "learning_rate": 0.0006, + "loss": 4.059067249298096, + "step": 7057 + }, + { + "epoch": 98.02794759825328, + "grad_norm": 0.013799279928207397, + "learning_rate": 0.0006, + "loss": 3.975419044494629, + "step": 7058 + }, + { + "epoch": 98.04192139737991, + "grad_norm": 0.014793423935770988, + "learning_rate": 0.0006, + "loss": 4.163051605224609, + "step": 7059 + }, + { + "epoch": 98.05589519650655, + "grad_norm": 0.015820559114217758, + "learning_rate": 0.0006, + "loss": 4.032471656799316, + "step": 7060 + }, + { + "epoch": 98.06986899563319, + "grad_norm": 0.014252204447984695, + "learning_rate": 0.0006, + "loss": 4.038850784301758, + "step": 7061 + }, + { + "epoch": 98.08384279475983, + "grad_norm": 0.01272337045520544, + "learning_rate": 0.0006, + "loss": 4.054554462432861, + "step": 7062 + }, + { + "epoch": 98.09781659388646, + "grad_norm": 0.01498495228588581, + "learning_rate": 0.0006, + "loss": 4.009291648864746, + "step": 7063 + }, + { + "epoch": 98.1117903930131, + "grad_norm": 0.015501809306442738, + "learning_rate": 0.0006, + "loss": 3.999412775039673, + "step": 7064 + }, + { + "epoch": 98.12576419213974, + "grad_norm": 0.015108918771147728, + "learning_rate": 0.0006, + "loss": 4.015168190002441, + "step": 7065 + }, + { + "epoch": 98.13973799126637, + "grad_norm": 0.01489692460745573, + "learning_rate": 0.0006, + "loss": 4.163906097412109, + "step": 7066 + }, + { + "epoch": 98.15371179039302, + "grad_norm": 0.014246871694922447, + "learning_rate": 0.0006, + "loss": 3.993173122406006, + "step": 7067 + }, + { + "epoch": 98.16768558951965, + "grad_norm": 0.015163703821599483, + "learning_rate": 0.0006, + "loss": 3.908207893371582, + "step": 7068 + }, + { + "epoch": 98.18165938864628, + "grad_norm": 0.0141439288854599, + "learning_rate": 0.0006, + "loss": 4.064990997314453, + "step": 7069 + }, + { + "epoch": 98.19563318777293, + "grad_norm": 0.013251710683107376, + "learning_rate": 0.0006, + "loss": 4.042545318603516, + "step": 7070 + }, + { + "epoch": 98.20960698689956, + "grad_norm": 0.01428674440830946, + "learning_rate": 0.0006, + "loss": 4.041232585906982, + "step": 7071 + }, + { + "epoch": 98.2235807860262, + "grad_norm": 0.012818179093301296, + "learning_rate": 0.0006, + "loss": 3.9160139560699463, + "step": 7072 + }, + { + "epoch": 98.23755458515284, + "grad_norm": 0.012890368700027466, + "learning_rate": 0.0006, + "loss": 3.9687235355377197, + "step": 7073 + }, + { + "epoch": 98.25152838427948, + "grad_norm": 0.014132961630821228, + "learning_rate": 0.0006, + "loss": 4.066800117492676, + "step": 7074 + }, + { + "epoch": 98.26550218340611, + "grad_norm": 0.013544383458793163, + "learning_rate": 0.0006, + "loss": 4.111777305603027, + "step": 7075 + }, + { + "epoch": 98.27947598253274, + "grad_norm": 0.014414685778319836, + "learning_rate": 0.0006, + "loss": 4.109633445739746, + "step": 7076 + }, + { + "epoch": 98.29344978165939, + "grad_norm": 0.015201127156615257, + "learning_rate": 0.0006, + "loss": 4.033087253570557, + "step": 7077 + }, + { + "epoch": 98.30742358078602, + "grad_norm": 0.016936076804995537, + "learning_rate": 0.0006, + "loss": 4.10897970199585, + "step": 7078 + }, + { + "epoch": 98.32139737991267, + "grad_norm": 0.01556481048464775, + "learning_rate": 0.0006, + "loss": 3.968942165374756, + "step": 7079 + }, + { + "epoch": 98.3353711790393, + "grad_norm": 0.015971971675753593, + "learning_rate": 0.0006, + "loss": 4.024805068969727, + "step": 7080 + }, + { + "epoch": 98.34934497816595, + "grad_norm": 0.016742313280701637, + "learning_rate": 0.0006, + "loss": 4.037862777709961, + "step": 7081 + }, + { + "epoch": 98.36331877729258, + "grad_norm": 0.015753688290715218, + "learning_rate": 0.0006, + "loss": 4.015701770782471, + "step": 7082 + }, + { + "epoch": 98.37729257641921, + "grad_norm": 0.014211287721991539, + "learning_rate": 0.0006, + "loss": 4.068171501159668, + "step": 7083 + }, + { + "epoch": 98.39126637554585, + "grad_norm": 0.013903875835239887, + "learning_rate": 0.0006, + "loss": 4.052516937255859, + "step": 7084 + }, + { + "epoch": 98.40524017467249, + "grad_norm": 0.015590954571962357, + "learning_rate": 0.0006, + "loss": 4.056125164031982, + "step": 7085 + }, + { + "epoch": 98.41921397379913, + "grad_norm": 0.016314011067152023, + "learning_rate": 0.0006, + "loss": 4.014211177825928, + "step": 7086 + }, + { + "epoch": 98.43318777292576, + "grad_norm": 0.017208732664585114, + "learning_rate": 0.0006, + "loss": 4.0615339279174805, + "step": 7087 + }, + { + "epoch": 98.44716157205241, + "grad_norm": 0.017213109880685806, + "learning_rate": 0.0006, + "loss": 3.9580445289611816, + "step": 7088 + }, + { + "epoch": 98.46113537117904, + "grad_norm": 0.016858434304594994, + "learning_rate": 0.0006, + "loss": 3.9932594299316406, + "step": 7089 + }, + { + "epoch": 98.47510917030567, + "grad_norm": 0.014381436631083488, + "learning_rate": 0.0006, + "loss": 4.159034729003906, + "step": 7090 + }, + { + "epoch": 98.48908296943232, + "grad_norm": 0.013713826425373554, + "learning_rate": 0.0006, + "loss": 4.055055141448975, + "step": 7091 + }, + { + "epoch": 98.50305676855895, + "grad_norm": 0.014764957129955292, + "learning_rate": 0.0006, + "loss": 4.059102535247803, + "step": 7092 + }, + { + "epoch": 98.5170305676856, + "grad_norm": 0.014522310346364975, + "learning_rate": 0.0006, + "loss": 4.093504905700684, + "step": 7093 + }, + { + "epoch": 98.53100436681223, + "grad_norm": 0.013484988361597061, + "learning_rate": 0.0006, + "loss": 4.079945087432861, + "step": 7094 + }, + { + "epoch": 98.54497816593886, + "grad_norm": 0.013231529854238033, + "learning_rate": 0.0006, + "loss": 4.046164512634277, + "step": 7095 + }, + { + "epoch": 98.5589519650655, + "grad_norm": 0.01392320916056633, + "learning_rate": 0.0006, + "loss": 3.975592613220215, + "step": 7096 + }, + { + "epoch": 98.57292576419214, + "grad_norm": 0.012323490343987942, + "learning_rate": 0.0006, + "loss": 3.9167721271514893, + "step": 7097 + }, + { + "epoch": 98.58689956331878, + "grad_norm": 0.012565829791128635, + "learning_rate": 0.0006, + "loss": 4.085109710693359, + "step": 7098 + }, + { + "epoch": 98.60087336244541, + "grad_norm": 0.013498359359800816, + "learning_rate": 0.0006, + "loss": 4.102974891662598, + "step": 7099 + }, + { + "epoch": 98.61484716157206, + "grad_norm": 0.014182298444211483, + "learning_rate": 0.0006, + "loss": 4.041601657867432, + "step": 7100 + }, + { + "epoch": 98.62882096069869, + "grad_norm": 0.013925517909228802, + "learning_rate": 0.0006, + "loss": 4.115788459777832, + "step": 7101 + }, + { + "epoch": 98.64279475982532, + "grad_norm": 0.03483536094427109, + "learning_rate": 0.0006, + "loss": 4.105683326721191, + "step": 7102 + }, + { + "epoch": 98.65676855895197, + "grad_norm": 0.014278477989137173, + "learning_rate": 0.0006, + "loss": 4.04194974899292, + "step": 7103 + }, + { + "epoch": 98.6707423580786, + "grad_norm": 0.015309838578104973, + "learning_rate": 0.0006, + "loss": 4.094045162200928, + "step": 7104 + }, + { + "epoch": 98.68471615720524, + "grad_norm": 0.015831857919692993, + "learning_rate": 0.0006, + "loss": 3.9989206790924072, + "step": 7105 + }, + { + "epoch": 98.69868995633188, + "grad_norm": 0.014071783982217312, + "learning_rate": 0.0006, + "loss": 4.0480852127075195, + "step": 7106 + }, + { + "epoch": 98.71266375545852, + "grad_norm": 0.014099884778261185, + "learning_rate": 0.0006, + "loss": 4.0519633293151855, + "step": 7107 + }, + { + "epoch": 98.72663755458515, + "grad_norm": 0.0143047496676445, + "learning_rate": 0.0006, + "loss": 3.9782276153564453, + "step": 7108 + }, + { + "epoch": 98.74061135371178, + "grad_norm": 0.013741862028837204, + "learning_rate": 0.0006, + "loss": 4.035103797912598, + "step": 7109 + }, + { + "epoch": 98.75458515283843, + "grad_norm": 0.014014648273587227, + "learning_rate": 0.0006, + "loss": 4.09281063079834, + "step": 7110 + }, + { + "epoch": 98.76855895196506, + "grad_norm": 0.013909446075558662, + "learning_rate": 0.0006, + "loss": 4.033784866333008, + "step": 7111 + }, + { + "epoch": 98.78253275109171, + "grad_norm": 0.013188562355935574, + "learning_rate": 0.0006, + "loss": 4.095016956329346, + "step": 7112 + }, + { + "epoch": 98.79650655021834, + "grad_norm": 0.013292612507939339, + "learning_rate": 0.0006, + "loss": 4.000730514526367, + "step": 7113 + }, + { + "epoch": 98.81048034934499, + "grad_norm": 0.013189040124416351, + "learning_rate": 0.0006, + "loss": 4.087883949279785, + "step": 7114 + }, + { + "epoch": 98.82445414847162, + "grad_norm": 0.013953711837530136, + "learning_rate": 0.0006, + "loss": 4.051388740539551, + "step": 7115 + }, + { + "epoch": 98.83842794759825, + "grad_norm": 0.014403220266103745, + "learning_rate": 0.0006, + "loss": 3.9786248207092285, + "step": 7116 + }, + { + "epoch": 98.8524017467249, + "grad_norm": 0.013747540302574635, + "learning_rate": 0.0006, + "loss": 3.9938254356384277, + "step": 7117 + }, + { + "epoch": 98.86637554585153, + "grad_norm": 0.014865634962916374, + "learning_rate": 0.0006, + "loss": 4.075680255889893, + "step": 7118 + }, + { + "epoch": 98.88034934497817, + "grad_norm": 0.015241672284901142, + "learning_rate": 0.0006, + "loss": 3.955733299255371, + "step": 7119 + }, + { + "epoch": 98.8943231441048, + "grad_norm": 0.013307510875165462, + "learning_rate": 0.0006, + "loss": 3.94956374168396, + "step": 7120 + }, + { + "epoch": 98.90829694323143, + "grad_norm": 0.013500694185495377, + "learning_rate": 0.0006, + "loss": 4.120021343231201, + "step": 7121 + }, + { + "epoch": 98.92227074235808, + "grad_norm": 0.015430806204676628, + "learning_rate": 0.0006, + "loss": 4.108362674713135, + "step": 7122 + }, + { + "epoch": 98.93624454148471, + "grad_norm": 0.015447405166924, + "learning_rate": 0.0006, + "loss": 4.040914058685303, + "step": 7123 + }, + { + "epoch": 98.95021834061136, + "grad_norm": 0.014040131121873856, + "learning_rate": 0.0006, + "loss": 4.086688995361328, + "step": 7124 + }, + { + "epoch": 98.96419213973799, + "grad_norm": 0.0136969443410635, + "learning_rate": 0.0006, + "loss": 4.083828926086426, + "step": 7125 + }, + { + "epoch": 98.97816593886463, + "grad_norm": 0.01413224171847105, + "learning_rate": 0.0006, + "loss": 4.142146110534668, + "step": 7126 + }, + { + "epoch": 98.99213973799127, + "grad_norm": 0.012932452373206615, + "learning_rate": 0.0006, + "loss": 4.1193342208862305, + "step": 7127 + }, + { + "epoch": 99.0, + "grad_norm": 0.01586783304810524, + "learning_rate": 0.0006, + "loss": 4.191825866699219, + "step": 7128 + }, + { + "epoch": 99.0, + "eval_loss": 4.550753116607666, + "eval_runtime": 56.4315, + "eval_samples_per_second": 43.274, + "eval_steps_per_second": 1.364, + "step": 7128 + }, + { + "epoch": 99.01397379912663, + "grad_norm": 0.016847660765051842, + "learning_rate": 0.0006, + "loss": 3.9874870777130127, + "step": 7129 + }, + { + "epoch": 99.02794759825328, + "grad_norm": 0.016426295042037964, + "learning_rate": 0.0006, + "loss": 4.108767032623291, + "step": 7130 + }, + { + "epoch": 99.04192139737991, + "grad_norm": 0.01687243953347206, + "learning_rate": 0.0006, + "loss": 3.9658308029174805, + "step": 7131 + }, + { + "epoch": 99.05589519650655, + "grad_norm": 0.016468465328216553, + "learning_rate": 0.0006, + "loss": 3.91381573677063, + "step": 7132 + }, + { + "epoch": 99.06986899563319, + "grad_norm": 0.013777337968349457, + "learning_rate": 0.0006, + "loss": 3.9960741996765137, + "step": 7133 + }, + { + "epoch": 99.08384279475983, + "grad_norm": 0.014061779715120792, + "learning_rate": 0.0006, + "loss": 4.069479942321777, + "step": 7134 + }, + { + "epoch": 99.09781659388646, + "grad_norm": 0.01513410173356533, + "learning_rate": 0.0006, + "loss": 4.145164489746094, + "step": 7135 + }, + { + "epoch": 99.1117903930131, + "grad_norm": 0.01567850261926651, + "learning_rate": 0.0006, + "loss": 4.016868591308594, + "step": 7136 + }, + { + "epoch": 99.12576419213974, + "grad_norm": 0.01662704534828663, + "learning_rate": 0.0006, + "loss": 3.966663360595703, + "step": 7137 + }, + { + "epoch": 99.13973799126637, + "grad_norm": 0.01690712384879589, + "learning_rate": 0.0006, + "loss": 4.003656387329102, + "step": 7138 + }, + { + "epoch": 99.15371179039302, + "grad_norm": 0.01541030965745449, + "learning_rate": 0.0006, + "loss": 3.9990856647491455, + "step": 7139 + }, + { + "epoch": 99.16768558951965, + "grad_norm": 0.014438044279813766, + "learning_rate": 0.0006, + "loss": 4.089632034301758, + "step": 7140 + }, + { + "epoch": 99.18165938864628, + "grad_norm": 0.01470417995005846, + "learning_rate": 0.0006, + "loss": 4.018681049346924, + "step": 7141 + }, + { + "epoch": 99.19563318777293, + "grad_norm": 0.015567616559565067, + "learning_rate": 0.0006, + "loss": 4.067337989807129, + "step": 7142 + }, + { + "epoch": 99.20960698689956, + "grad_norm": 0.01784040965139866, + "learning_rate": 0.0006, + "loss": 4.083293914794922, + "step": 7143 + }, + { + "epoch": 99.2235807860262, + "grad_norm": 0.017351210117340088, + "learning_rate": 0.0006, + "loss": 4.115224361419678, + "step": 7144 + }, + { + "epoch": 99.23755458515284, + "grad_norm": 0.014561209827661514, + "learning_rate": 0.0006, + "loss": 4.1001057624816895, + "step": 7145 + }, + { + "epoch": 99.25152838427948, + "grad_norm": 0.013986383564770222, + "learning_rate": 0.0006, + "loss": 3.9758031368255615, + "step": 7146 + }, + { + "epoch": 99.26550218340611, + "grad_norm": 0.014053978025913239, + "learning_rate": 0.0006, + "loss": 4.1420369148254395, + "step": 7147 + }, + { + "epoch": 99.27947598253274, + "grad_norm": 0.014878278598189354, + "learning_rate": 0.0006, + "loss": 4.018189430236816, + "step": 7148 + }, + { + "epoch": 99.29344978165939, + "grad_norm": 0.014634845778346062, + "learning_rate": 0.0006, + "loss": 4.062045574188232, + "step": 7149 + }, + { + "epoch": 99.30742358078602, + "grad_norm": 0.014713970944285393, + "learning_rate": 0.0006, + "loss": 4.017163276672363, + "step": 7150 + }, + { + "epoch": 99.32139737991267, + "grad_norm": 0.01589985005557537, + "learning_rate": 0.0006, + "loss": 4.093014717102051, + "step": 7151 + }, + { + "epoch": 99.3353711790393, + "grad_norm": 0.016232378780841827, + "learning_rate": 0.0006, + "loss": 4.029753684997559, + "step": 7152 + }, + { + "epoch": 99.34934497816595, + "grad_norm": 0.016340898349881172, + "learning_rate": 0.0006, + "loss": 4.056434631347656, + "step": 7153 + }, + { + "epoch": 99.36331877729258, + "grad_norm": 0.01587427407503128, + "learning_rate": 0.0006, + "loss": 3.932020425796509, + "step": 7154 + }, + { + "epoch": 99.37729257641921, + "grad_norm": 0.014977093786001205, + "learning_rate": 0.0006, + "loss": 4.015125751495361, + "step": 7155 + }, + { + "epoch": 99.39126637554585, + "grad_norm": 0.014813411049544811, + "learning_rate": 0.0006, + "loss": 4.088035583496094, + "step": 7156 + }, + { + "epoch": 99.40524017467249, + "grad_norm": 0.014407774433493614, + "learning_rate": 0.0006, + "loss": 3.9979054927825928, + "step": 7157 + }, + { + "epoch": 99.41921397379913, + "grad_norm": 0.013969994150102139, + "learning_rate": 0.0006, + "loss": 4.180194854736328, + "step": 7158 + }, + { + "epoch": 99.43318777292576, + "grad_norm": 0.0159455519169569, + "learning_rate": 0.0006, + "loss": 4.049341201782227, + "step": 7159 + }, + { + "epoch": 99.44716157205241, + "grad_norm": 0.016426580026745796, + "learning_rate": 0.0006, + "loss": 4.01624059677124, + "step": 7160 + }, + { + "epoch": 99.46113537117904, + "grad_norm": 0.014921349473297596, + "learning_rate": 0.0006, + "loss": 4.003138542175293, + "step": 7161 + }, + { + "epoch": 99.47510917030567, + "grad_norm": 0.014607159420847893, + "learning_rate": 0.0006, + "loss": 3.993896484375, + "step": 7162 + }, + { + "epoch": 99.48908296943232, + "grad_norm": 0.014638705179095268, + "learning_rate": 0.0006, + "loss": 4.015478134155273, + "step": 7163 + }, + { + "epoch": 99.50305676855895, + "grad_norm": 0.01346661988645792, + "learning_rate": 0.0006, + "loss": 3.9999020099639893, + "step": 7164 + }, + { + "epoch": 99.5170305676856, + "grad_norm": 0.013515726663172245, + "learning_rate": 0.0006, + "loss": 4.045872688293457, + "step": 7165 + }, + { + "epoch": 99.53100436681223, + "grad_norm": 0.013773965649306774, + "learning_rate": 0.0006, + "loss": 4.094198703765869, + "step": 7166 + }, + { + "epoch": 99.54497816593886, + "grad_norm": 0.014282289892435074, + "learning_rate": 0.0006, + "loss": 4.091470718383789, + "step": 7167 + }, + { + "epoch": 99.5589519650655, + "grad_norm": 0.015272347256541252, + "learning_rate": 0.0006, + "loss": 4.026535987854004, + "step": 7168 + }, + { + "epoch": 99.57292576419214, + "grad_norm": 0.015618950128555298, + "learning_rate": 0.0006, + "loss": 4.006448745727539, + "step": 7169 + }, + { + "epoch": 99.58689956331878, + "grad_norm": 0.015366379171609879, + "learning_rate": 0.0006, + "loss": 4.025970458984375, + "step": 7170 + }, + { + "epoch": 99.60087336244541, + "grad_norm": 0.013261756859719753, + "learning_rate": 0.0006, + "loss": 3.980401039123535, + "step": 7171 + }, + { + "epoch": 99.61484716157206, + "grad_norm": 0.013438165187835693, + "learning_rate": 0.0006, + "loss": 4.0106096267700195, + "step": 7172 + }, + { + "epoch": 99.62882096069869, + "grad_norm": 0.015591824427247047, + "learning_rate": 0.0006, + "loss": 4.0458784103393555, + "step": 7173 + }, + { + "epoch": 99.64279475982532, + "grad_norm": 0.014863507822155952, + "learning_rate": 0.0006, + "loss": 4.014740943908691, + "step": 7174 + }, + { + "epoch": 99.65676855895197, + "grad_norm": 0.012508790008723736, + "learning_rate": 0.0006, + "loss": 4.022995948791504, + "step": 7175 + }, + { + "epoch": 99.6707423580786, + "grad_norm": 0.013107656501233578, + "learning_rate": 0.0006, + "loss": 4.085544109344482, + "step": 7176 + }, + { + "epoch": 99.68471615720524, + "grad_norm": 0.013550374656915665, + "learning_rate": 0.0006, + "loss": 4.1239237785339355, + "step": 7177 + }, + { + "epoch": 99.69868995633188, + "grad_norm": 0.012947923503816128, + "learning_rate": 0.0006, + "loss": 4.01287317276001, + "step": 7178 + }, + { + "epoch": 99.71266375545852, + "grad_norm": 0.015236291103065014, + "learning_rate": 0.0006, + "loss": 4.075140476226807, + "step": 7179 + }, + { + "epoch": 99.72663755458515, + "grad_norm": 0.014862542040646076, + "learning_rate": 0.0006, + "loss": 4.111439228057861, + "step": 7180 + }, + { + "epoch": 99.74061135371178, + "grad_norm": 0.013883531093597412, + "learning_rate": 0.0006, + "loss": 3.9860856533050537, + "step": 7181 + }, + { + "epoch": 99.75458515283843, + "grad_norm": 0.01507870014756918, + "learning_rate": 0.0006, + "loss": 4.096504211425781, + "step": 7182 + }, + { + "epoch": 99.76855895196506, + "grad_norm": 0.015378627926111221, + "learning_rate": 0.0006, + "loss": 4.034113883972168, + "step": 7183 + }, + { + "epoch": 99.78253275109171, + "grad_norm": 0.013696327805519104, + "learning_rate": 0.0006, + "loss": 4.026954650878906, + "step": 7184 + }, + { + "epoch": 99.79650655021834, + "grad_norm": 0.013519738800823689, + "learning_rate": 0.0006, + "loss": 4.040035247802734, + "step": 7185 + }, + { + "epoch": 99.81048034934499, + "grad_norm": 0.015677500516176224, + "learning_rate": 0.0006, + "loss": 4.057739734649658, + "step": 7186 + }, + { + "epoch": 99.82445414847162, + "grad_norm": 0.014739335514605045, + "learning_rate": 0.0006, + "loss": 4.022522449493408, + "step": 7187 + }, + { + "epoch": 99.83842794759825, + "grad_norm": 0.014079970307648182, + "learning_rate": 0.0006, + "loss": 3.9286317825317383, + "step": 7188 + }, + { + "epoch": 99.8524017467249, + "grad_norm": 0.014899874106049538, + "learning_rate": 0.0006, + "loss": 4.081925868988037, + "step": 7189 + }, + { + "epoch": 99.86637554585153, + "grad_norm": 0.014767617918550968, + "learning_rate": 0.0006, + "loss": 4.029382705688477, + "step": 7190 + }, + { + "epoch": 99.88034934497817, + "grad_norm": 0.014914393424987793, + "learning_rate": 0.0006, + "loss": 4.056460857391357, + "step": 7191 + }, + { + "epoch": 99.8943231441048, + "grad_norm": 0.01720450073480606, + "learning_rate": 0.0006, + "loss": 4.0184478759765625, + "step": 7192 + }, + { + "epoch": 99.90829694323143, + "grad_norm": 0.016545087099075317, + "learning_rate": 0.0006, + "loss": 3.9638831615448, + "step": 7193 + }, + { + "epoch": 99.92227074235808, + "grad_norm": 0.014489716850221157, + "learning_rate": 0.0006, + "loss": 4.0267653465271, + "step": 7194 + }, + { + "epoch": 99.93624454148471, + "grad_norm": 0.01438206061720848, + "learning_rate": 0.0006, + "loss": 4.212850570678711, + "step": 7195 + }, + { + "epoch": 99.95021834061136, + "grad_norm": 0.017636865377426147, + "learning_rate": 0.0006, + "loss": 3.942308187484741, + "step": 7196 + }, + { + "epoch": 99.96419213973799, + "grad_norm": 0.019075388088822365, + "learning_rate": 0.0006, + "loss": 4.026660919189453, + "step": 7197 + }, + { + "epoch": 99.97816593886463, + "grad_norm": 0.016569023951888084, + "learning_rate": 0.0006, + "loss": 4.018840789794922, + "step": 7198 + }, + { + "epoch": 99.99213973799127, + "grad_norm": 0.014842798933386803, + "learning_rate": 0.0006, + "loss": 4.041904449462891, + "step": 7199 + }, + { + "epoch": 100.0, + "grad_norm": 0.0167684406042099, + "learning_rate": 0.0006, + "loss": 3.9079346656799316, + "step": 7200 + }, + { + "epoch": 100.0, + "eval_loss": 4.585540771484375, + "eval_runtime": 56.5095, + "eval_samples_per_second": 43.214, + "eval_steps_per_second": 1.363, + "step": 7200 + }, + { + "epoch": 100.01397379912663, + "grad_norm": 0.016783876344561577, + "learning_rate": 0.0006, + "loss": 3.998840808868408, + "step": 7201 + }, + { + "epoch": 100.02794759825328, + "grad_norm": 0.015306975692510605, + "learning_rate": 0.0006, + "loss": 4.161236763000488, + "step": 7202 + }, + { + "epoch": 100.04192139737991, + "grad_norm": 0.016536490991711617, + "learning_rate": 0.0006, + "loss": 4.044086933135986, + "step": 7203 + }, + { + "epoch": 100.05589519650655, + "grad_norm": 0.016115620732307434, + "learning_rate": 0.0006, + "loss": 3.9210472106933594, + "step": 7204 + }, + { + "epoch": 100.06986899563319, + "grad_norm": 0.014647277072072029, + "learning_rate": 0.0006, + "loss": 3.9738388061523438, + "step": 7205 + }, + { + "epoch": 100.08384279475983, + "grad_norm": 0.013562940992414951, + "learning_rate": 0.0006, + "loss": 4.093536376953125, + "step": 7206 + }, + { + "epoch": 100.09781659388646, + "grad_norm": 0.013523460365831852, + "learning_rate": 0.0006, + "loss": 3.985286235809326, + "step": 7207 + }, + { + "epoch": 100.1117903930131, + "grad_norm": 0.01463227067142725, + "learning_rate": 0.0006, + "loss": 3.8212950229644775, + "step": 7208 + }, + { + "epoch": 100.12576419213974, + "grad_norm": 0.015067536383867264, + "learning_rate": 0.0006, + "loss": 4.050893783569336, + "step": 7209 + }, + { + "epoch": 100.13973799126637, + "grad_norm": 0.015035216696560383, + "learning_rate": 0.0006, + "loss": 4.108796119689941, + "step": 7210 + }, + { + "epoch": 100.15371179039302, + "grad_norm": 0.01694556325674057, + "learning_rate": 0.0006, + "loss": 4.001124382019043, + "step": 7211 + }, + { + "epoch": 100.16768558951965, + "grad_norm": 0.01810646802186966, + "learning_rate": 0.0006, + "loss": 4.122286319732666, + "step": 7212 + }, + { + "epoch": 100.18165938864628, + "grad_norm": 0.019156092777848244, + "learning_rate": 0.0006, + "loss": 4.087690353393555, + "step": 7213 + }, + { + "epoch": 100.19563318777293, + "grad_norm": 0.018056577071547508, + "learning_rate": 0.0006, + "loss": 4.135015487670898, + "step": 7214 + }, + { + "epoch": 100.20960698689956, + "grad_norm": 0.017466023564338684, + "learning_rate": 0.0006, + "loss": 4.028027534484863, + "step": 7215 + }, + { + "epoch": 100.2235807860262, + "grad_norm": 0.017239956185221672, + "learning_rate": 0.0006, + "loss": 4.128888130187988, + "step": 7216 + }, + { + "epoch": 100.23755458515284, + "grad_norm": 0.016968458890914917, + "learning_rate": 0.0006, + "loss": 3.998007297515869, + "step": 7217 + }, + { + "epoch": 100.25152838427948, + "grad_norm": 0.016718817874789238, + "learning_rate": 0.0006, + "loss": 4.125210762023926, + "step": 7218 + }, + { + "epoch": 100.26550218340611, + "grad_norm": 0.0164450965821743, + "learning_rate": 0.0006, + "loss": 4.000813007354736, + "step": 7219 + }, + { + "epoch": 100.27947598253274, + "grad_norm": 0.01655936799943447, + "learning_rate": 0.0006, + "loss": 3.9809513092041016, + "step": 7220 + }, + { + "epoch": 100.29344978165939, + "grad_norm": 0.015745731070637703, + "learning_rate": 0.0006, + "loss": 4.089069366455078, + "step": 7221 + }, + { + "epoch": 100.30742358078602, + "grad_norm": 0.01796283759176731, + "learning_rate": 0.0006, + "loss": 4.049221992492676, + "step": 7222 + }, + { + "epoch": 100.32139737991267, + "grad_norm": 0.018306732177734375, + "learning_rate": 0.0006, + "loss": 4.043508052825928, + "step": 7223 + }, + { + "epoch": 100.3353711790393, + "grad_norm": 0.016871308907866478, + "learning_rate": 0.0006, + "loss": 4.051634311676025, + "step": 7224 + }, + { + "epoch": 100.34934497816595, + "grad_norm": 0.016082527115941048, + "learning_rate": 0.0006, + "loss": 3.905829668045044, + "step": 7225 + }, + { + "epoch": 100.36331877729258, + "grad_norm": 0.01602019928395748, + "learning_rate": 0.0006, + "loss": 4.039409637451172, + "step": 7226 + }, + { + "epoch": 100.37729257641921, + "grad_norm": 0.01674795150756836, + "learning_rate": 0.0006, + "loss": 3.9757080078125, + "step": 7227 + }, + { + "epoch": 100.39126637554585, + "grad_norm": 0.014572949148714542, + "learning_rate": 0.0006, + "loss": 4.073253631591797, + "step": 7228 + }, + { + "epoch": 100.40524017467249, + "grad_norm": 0.015504667535424232, + "learning_rate": 0.0006, + "loss": 4.122897624969482, + "step": 7229 + }, + { + "epoch": 100.41921397379913, + "grad_norm": 0.0163279939442873, + "learning_rate": 0.0006, + "loss": 4.1488823890686035, + "step": 7230 + }, + { + "epoch": 100.43318777292576, + "grad_norm": 0.016203004866838455, + "learning_rate": 0.0006, + "loss": 4.099796295166016, + "step": 7231 + }, + { + "epoch": 100.44716157205241, + "grad_norm": 0.014418032020330429, + "learning_rate": 0.0006, + "loss": 3.994591474533081, + "step": 7232 + }, + { + "epoch": 100.46113537117904, + "grad_norm": 0.014775440096855164, + "learning_rate": 0.0006, + "loss": 4.07340145111084, + "step": 7233 + }, + { + "epoch": 100.47510917030567, + "grad_norm": 0.015920625999569893, + "learning_rate": 0.0006, + "loss": 3.985908269882202, + "step": 7234 + }, + { + "epoch": 100.48908296943232, + "grad_norm": 0.015350693836808205, + "learning_rate": 0.0006, + "loss": 4.007932662963867, + "step": 7235 + }, + { + "epoch": 100.50305676855895, + "grad_norm": 0.014701326377689838, + "learning_rate": 0.0006, + "loss": 4.151021957397461, + "step": 7236 + }, + { + "epoch": 100.5170305676856, + "grad_norm": 0.013621610589325428, + "learning_rate": 0.0006, + "loss": 3.988795280456543, + "step": 7237 + }, + { + "epoch": 100.53100436681223, + "grad_norm": 0.014496153220534325, + "learning_rate": 0.0006, + "loss": 4.019025802612305, + "step": 7238 + }, + { + "epoch": 100.54497816593886, + "grad_norm": 0.013380124233663082, + "learning_rate": 0.0006, + "loss": 4.090785980224609, + "step": 7239 + }, + { + "epoch": 100.5589519650655, + "grad_norm": 0.01378310564905405, + "learning_rate": 0.0006, + "loss": 4.158342361450195, + "step": 7240 + }, + { + "epoch": 100.57292576419214, + "grad_norm": 0.01402236521244049, + "learning_rate": 0.0006, + "loss": 4.108238220214844, + "step": 7241 + }, + { + "epoch": 100.58689956331878, + "grad_norm": 0.014660182408988476, + "learning_rate": 0.0006, + "loss": 4.052890300750732, + "step": 7242 + }, + { + "epoch": 100.60087336244541, + "grad_norm": 0.01419265940785408, + "learning_rate": 0.0006, + "loss": 4.048383712768555, + "step": 7243 + }, + { + "epoch": 100.61484716157206, + "grad_norm": 0.014536422677338123, + "learning_rate": 0.0006, + "loss": 3.938699960708618, + "step": 7244 + }, + { + "epoch": 100.62882096069869, + "grad_norm": 0.013167395256459713, + "learning_rate": 0.0006, + "loss": 4.13656759262085, + "step": 7245 + }, + { + "epoch": 100.64279475982532, + "grad_norm": 0.013563545420765877, + "learning_rate": 0.0006, + "loss": 4.013095855712891, + "step": 7246 + }, + { + "epoch": 100.65676855895197, + "grad_norm": 0.013953451067209244, + "learning_rate": 0.0006, + "loss": 4.091291427612305, + "step": 7247 + }, + { + "epoch": 100.6707423580786, + "grad_norm": 0.014537926763296127, + "learning_rate": 0.0006, + "loss": 4.041815757751465, + "step": 7248 + }, + { + "epoch": 100.68471615720524, + "grad_norm": 0.014321094378829002, + "learning_rate": 0.0006, + "loss": 4.033531665802002, + "step": 7249 + }, + { + "epoch": 100.69868995633188, + "grad_norm": 0.013944404199719429, + "learning_rate": 0.0006, + "loss": 4.119833946228027, + "step": 7250 + }, + { + "epoch": 100.71266375545852, + "grad_norm": 0.01371681783348322, + "learning_rate": 0.0006, + "loss": 3.965501546859741, + "step": 7251 + }, + { + "epoch": 100.72663755458515, + "grad_norm": 0.012983391061425209, + "learning_rate": 0.0006, + "loss": 3.987842559814453, + "step": 7252 + }, + { + "epoch": 100.74061135371178, + "grad_norm": 0.013693883083760738, + "learning_rate": 0.0006, + "loss": 4.066164493560791, + "step": 7253 + }, + { + "epoch": 100.75458515283843, + "grad_norm": 0.014391372911632061, + "learning_rate": 0.0006, + "loss": 4.095022201538086, + "step": 7254 + }, + { + "epoch": 100.76855895196506, + "grad_norm": 0.013853080570697784, + "learning_rate": 0.0006, + "loss": 4.099021911621094, + "step": 7255 + }, + { + "epoch": 100.78253275109171, + "grad_norm": 0.014468333683907986, + "learning_rate": 0.0006, + "loss": 4.040180206298828, + "step": 7256 + }, + { + "epoch": 100.79650655021834, + "grad_norm": 0.015436802990734577, + "learning_rate": 0.0006, + "loss": 4.0907087326049805, + "step": 7257 + }, + { + "epoch": 100.81048034934499, + "grad_norm": 0.014863028191030025, + "learning_rate": 0.0006, + "loss": 4.075537204742432, + "step": 7258 + }, + { + "epoch": 100.82445414847162, + "grad_norm": 0.014109591953456402, + "learning_rate": 0.0006, + "loss": 3.922210693359375, + "step": 7259 + }, + { + "epoch": 100.83842794759825, + "grad_norm": 0.014673606492578983, + "learning_rate": 0.0006, + "loss": 4.013545036315918, + "step": 7260 + }, + { + "epoch": 100.8524017467249, + "grad_norm": 0.01422286219894886, + "learning_rate": 0.0006, + "loss": 3.926832675933838, + "step": 7261 + }, + { + "epoch": 100.86637554585153, + "grad_norm": 0.013322828337550163, + "learning_rate": 0.0006, + "loss": 4.032454490661621, + "step": 7262 + }, + { + "epoch": 100.88034934497817, + "grad_norm": 0.014322307892143726, + "learning_rate": 0.0006, + "loss": 4.034879684448242, + "step": 7263 + }, + { + "epoch": 100.8943231441048, + "grad_norm": 0.013952548615634441, + "learning_rate": 0.0006, + "loss": 4.092174530029297, + "step": 7264 + }, + { + "epoch": 100.90829694323143, + "grad_norm": 0.01494069304317236, + "learning_rate": 0.0006, + "loss": 4.051094055175781, + "step": 7265 + }, + { + "epoch": 100.92227074235808, + "grad_norm": 0.01506600808352232, + "learning_rate": 0.0006, + "loss": 4.150339603424072, + "step": 7266 + }, + { + "epoch": 100.93624454148471, + "grad_norm": 0.015410789288580418, + "learning_rate": 0.0006, + "loss": 4.029681205749512, + "step": 7267 + }, + { + "epoch": 100.95021834061136, + "grad_norm": 0.01526960451155901, + "learning_rate": 0.0006, + "loss": 3.9946770668029785, + "step": 7268 + }, + { + "epoch": 100.96419213973799, + "grad_norm": 0.01581135205924511, + "learning_rate": 0.0006, + "loss": 4.050585746765137, + "step": 7269 + }, + { + "epoch": 100.97816593886463, + "grad_norm": 0.015924792736768723, + "learning_rate": 0.0006, + "loss": 4.005122184753418, + "step": 7270 + }, + { + "epoch": 100.99213973799127, + "grad_norm": 0.01638958603143692, + "learning_rate": 0.0006, + "loss": 4.074165344238281, + "step": 7271 + }, + { + "epoch": 101.0, + "grad_norm": 0.016363589093089104, + "learning_rate": 0.0006, + "loss": 4.004770278930664, + "step": 7272 + }, + { + "epoch": 101.0, + "eval_loss": 4.521492958068848, + "eval_runtime": 56.7059, + "eval_samples_per_second": 43.064, + "eval_steps_per_second": 1.358, + "step": 7272 + }, + { + "epoch": 101.01397379912663, + "grad_norm": 0.013911865651607513, + "learning_rate": 0.0006, + "loss": 4.073930263519287, + "step": 7273 + }, + { + "epoch": 101.02794759825328, + "grad_norm": 0.014139825478196144, + "learning_rate": 0.0006, + "loss": 4.0691142082214355, + "step": 7274 + }, + { + "epoch": 101.04192139737991, + "grad_norm": 0.015765748918056488, + "learning_rate": 0.0006, + "loss": 4.112308502197266, + "step": 7275 + }, + { + "epoch": 101.05589519650655, + "grad_norm": 0.016765642911195755, + "learning_rate": 0.0006, + "loss": 3.9341437816619873, + "step": 7276 + }, + { + "epoch": 101.06986899563319, + "grad_norm": 0.016253530979156494, + "learning_rate": 0.0006, + "loss": 3.9987592697143555, + "step": 7277 + }, + { + "epoch": 101.08384279475983, + "grad_norm": 0.015670686960220337, + "learning_rate": 0.0006, + "loss": 4.0460357666015625, + "step": 7278 + }, + { + "epoch": 101.09781659388646, + "grad_norm": 0.01603814959526062, + "learning_rate": 0.0006, + "loss": 4.005774974822998, + "step": 7279 + }, + { + "epoch": 101.1117903930131, + "grad_norm": 0.016335027292370796, + "learning_rate": 0.0006, + "loss": 3.9924376010894775, + "step": 7280 + }, + { + "epoch": 101.12576419213974, + "grad_norm": 0.017239267006516457, + "learning_rate": 0.0006, + "loss": 4.034348487854004, + "step": 7281 + }, + { + "epoch": 101.13973799126637, + "grad_norm": 0.016026441007852554, + "learning_rate": 0.0006, + "loss": 4.054862022399902, + "step": 7282 + }, + { + "epoch": 101.15371179039302, + "grad_norm": 0.015502813272178173, + "learning_rate": 0.0006, + "loss": 3.974360466003418, + "step": 7283 + }, + { + "epoch": 101.16768558951965, + "grad_norm": 0.01737619936466217, + "learning_rate": 0.0006, + "loss": 3.991811752319336, + "step": 7284 + }, + { + "epoch": 101.18165938864628, + "grad_norm": 0.01761457324028015, + "learning_rate": 0.0006, + "loss": 4.021451473236084, + "step": 7285 + }, + { + "epoch": 101.19563318777293, + "grad_norm": 0.017094692215323448, + "learning_rate": 0.0006, + "loss": 4.00285530090332, + "step": 7286 + }, + { + "epoch": 101.20960698689956, + "grad_norm": 0.0163987148553133, + "learning_rate": 0.0006, + "loss": 4.061247825622559, + "step": 7287 + }, + { + "epoch": 101.2235807860262, + "grad_norm": 0.015596904791891575, + "learning_rate": 0.0006, + "loss": 3.971940040588379, + "step": 7288 + }, + { + "epoch": 101.23755458515284, + "grad_norm": 0.015181072056293488, + "learning_rate": 0.0006, + "loss": 4.127270698547363, + "step": 7289 + }, + { + "epoch": 101.25152838427948, + "grad_norm": 0.014719299040734768, + "learning_rate": 0.0006, + "loss": 4.095149993896484, + "step": 7290 + }, + { + "epoch": 101.26550218340611, + "grad_norm": 0.01476318296045065, + "learning_rate": 0.0006, + "loss": 4.091351509094238, + "step": 7291 + }, + { + "epoch": 101.27947598253274, + "grad_norm": 0.014008177444338799, + "learning_rate": 0.0006, + "loss": 3.9897727966308594, + "step": 7292 + }, + { + "epoch": 101.29344978165939, + "grad_norm": 0.012773942202329636, + "learning_rate": 0.0006, + "loss": 3.986448287963867, + "step": 7293 + }, + { + "epoch": 101.30742358078602, + "grad_norm": 0.013246177695691586, + "learning_rate": 0.0006, + "loss": 4.105729579925537, + "step": 7294 + }, + { + "epoch": 101.32139737991267, + "grad_norm": 0.013527474366128445, + "learning_rate": 0.0006, + "loss": 4.012033939361572, + "step": 7295 + }, + { + "epoch": 101.3353711790393, + "grad_norm": 0.013510984368622303, + "learning_rate": 0.0006, + "loss": 3.9836957454681396, + "step": 7296 + }, + { + "epoch": 101.34934497816595, + "grad_norm": 0.01403056737035513, + "learning_rate": 0.0006, + "loss": 4.100725173950195, + "step": 7297 + }, + { + "epoch": 101.36331877729258, + "grad_norm": 0.013127653859555721, + "learning_rate": 0.0006, + "loss": 3.977855682373047, + "step": 7298 + }, + { + "epoch": 101.37729257641921, + "grad_norm": 0.015479992143809795, + "learning_rate": 0.0006, + "loss": 3.97141170501709, + "step": 7299 + }, + { + "epoch": 101.39126637554585, + "grad_norm": 0.01583872176706791, + "learning_rate": 0.0006, + "loss": 4.0838518142700195, + "step": 7300 + }, + { + "epoch": 101.40524017467249, + "grad_norm": 0.015457033179700375, + "learning_rate": 0.0006, + "loss": 4.061662673950195, + "step": 7301 + }, + { + "epoch": 101.41921397379913, + "grad_norm": 0.01576954312622547, + "learning_rate": 0.0006, + "loss": 4.069742202758789, + "step": 7302 + }, + { + "epoch": 101.43318777292576, + "grad_norm": 0.015752265229821205, + "learning_rate": 0.0006, + "loss": 3.8633975982666016, + "step": 7303 + }, + { + "epoch": 101.44716157205241, + "grad_norm": 0.015041268430650234, + "learning_rate": 0.0006, + "loss": 4.118044853210449, + "step": 7304 + }, + { + "epoch": 101.46113537117904, + "grad_norm": 0.015786072239279747, + "learning_rate": 0.0006, + "loss": 4.060590744018555, + "step": 7305 + }, + { + "epoch": 101.47510917030567, + "grad_norm": 0.01664225198328495, + "learning_rate": 0.0006, + "loss": 3.9695353507995605, + "step": 7306 + }, + { + "epoch": 101.48908296943232, + "grad_norm": 0.016616834327578545, + "learning_rate": 0.0006, + "loss": 4.1170454025268555, + "step": 7307 + }, + { + "epoch": 101.50305676855895, + "grad_norm": 0.01654205098748207, + "learning_rate": 0.0006, + "loss": 3.999567985534668, + "step": 7308 + }, + { + "epoch": 101.5170305676856, + "grad_norm": 0.01749691553413868, + "learning_rate": 0.0006, + "loss": 4.1020917892456055, + "step": 7309 + }, + { + "epoch": 101.53100436681223, + "grad_norm": 0.01645602658390999, + "learning_rate": 0.0006, + "loss": 4.070579528808594, + "step": 7310 + }, + { + "epoch": 101.54497816593886, + "grad_norm": 0.015536080114543438, + "learning_rate": 0.0006, + "loss": 4.013237953186035, + "step": 7311 + }, + { + "epoch": 101.5589519650655, + "grad_norm": 0.016780823469161987, + "learning_rate": 0.0006, + "loss": 4.027227401733398, + "step": 7312 + }, + { + "epoch": 101.57292576419214, + "grad_norm": 0.017896618694067, + "learning_rate": 0.0006, + "loss": 3.95709228515625, + "step": 7313 + }, + { + "epoch": 101.58689956331878, + "grad_norm": 0.01725279912352562, + "learning_rate": 0.0006, + "loss": 3.9525325298309326, + "step": 7314 + }, + { + "epoch": 101.60087336244541, + "grad_norm": 0.015343503095209599, + "learning_rate": 0.0006, + "loss": 4.032679557800293, + "step": 7315 + }, + { + "epoch": 101.61484716157206, + "grad_norm": 0.015583384782075882, + "learning_rate": 0.0006, + "loss": 3.9139232635498047, + "step": 7316 + }, + { + "epoch": 101.62882096069869, + "grad_norm": 0.01616024412214756, + "learning_rate": 0.0006, + "loss": 3.984429121017456, + "step": 7317 + }, + { + "epoch": 101.64279475982532, + "grad_norm": 0.015488795936107635, + "learning_rate": 0.0006, + "loss": 4.001983165740967, + "step": 7318 + }, + { + "epoch": 101.65676855895197, + "grad_norm": 0.015609036199748516, + "learning_rate": 0.0006, + "loss": 4.029224395751953, + "step": 7319 + }, + { + "epoch": 101.6707423580786, + "grad_norm": 0.01512704137712717, + "learning_rate": 0.0006, + "loss": 4.138855934143066, + "step": 7320 + }, + { + "epoch": 101.68471615720524, + "grad_norm": 0.014452820643782616, + "learning_rate": 0.0006, + "loss": 4.014325141906738, + "step": 7321 + }, + { + "epoch": 101.69868995633188, + "grad_norm": 0.013438846915960312, + "learning_rate": 0.0006, + "loss": 4.038142204284668, + "step": 7322 + }, + { + "epoch": 101.71266375545852, + "grad_norm": 0.014854568056762218, + "learning_rate": 0.0006, + "loss": 4.005454063415527, + "step": 7323 + }, + { + "epoch": 101.72663755458515, + "grad_norm": 0.01406457182019949, + "learning_rate": 0.0006, + "loss": 4.163865089416504, + "step": 7324 + }, + { + "epoch": 101.74061135371178, + "grad_norm": 0.01440865732729435, + "learning_rate": 0.0006, + "loss": 4.050951957702637, + "step": 7325 + }, + { + "epoch": 101.75458515283843, + "grad_norm": 0.013415827415883541, + "learning_rate": 0.0006, + "loss": 4.061484336853027, + "step": 7326 + }, + { + "epoch": 101.76855895196506, + "grad_norm": 0.012345070950686932, + "learning_rate": 0.0006, + "loss": 4.113116264343262, + "step": 7327 + }, + { + "epoch": 101.78253275109171, + "grad_norm": 0.01455951202660799, + "learning_rate": 0.0006, + "loss": 4.1283650398254395, + "step": 7328 + }, + { + "epoch": 101.79650655021834, + "grad_norm": 0.015864135697484016, + "learning_rate": 0.0006, + "loss": 4.084014892578125, + "step": 7329 + }, + { + "epoch": 101.81048034934499, + "grad_norm": 0.015105332247912884, + "learning_rate": 0.0006, + "loss": 4.136471748352051, + "step": 7330 + }, + { + "epoch": 101.82445414847162, + "grad_norm": 0.014927559532225132, + "learning_rate": 0.0006, + "loss": 4.039514541625977, + "step": 7331 + }, + { + "epoch": 101.83842794759825, + "grad_norm": 0.014508268795907497, + "learning_rate": 0.0006, + "loss": 4.052594184875488, + "step": 7332 + }, + { + "epoch": 101.8524017467249, + "grad_norm": 0.01513114757835865, + "learning_rate": 0.0006, + "loss": 4.084187984466553, + "step": 7333 + }, + { + "epoch": 101.86637554585153, + "grad_norm": 0.014415159821510315, + "learning_rate": 0.0006, + "loss": 3.8839468955993652, + "step": 7334 + }, + { + "epoch": 101.88034934497817, + "grad_norm": 0.013508095405995846, + "learning_rate": 0.0006, + "loss": 4.091848850250244, + "step": 7335 + }, + { + "epoch": 101.8943231441048, + "grad_norm": 0.013396956957876682, + "learning_rate": 0.0006, + "loss": 3.983945369720459, + "step": 7336 + }, + { + "epoch": 101.90829694323143, + "grad_norm": 0.013772518374025822, + "learning_rate": 0.0006, + "loss": 4.029302597045898, + "step": 7337 + }, + { + "epoch": 101.92227074235808, + "grad_norm": 0.01479576900601387, + "learning_rate": 0.0006, + "loss": 3.8791067600250244, + "step": 7338 + }, + { + "epoch": 101.93624454148471, + "grad_norm": 0.014177192002534866, + "learning_rate": 0.0006, + "loss": 4.044843673706055, + "step": 7339 + }, + { + "epoch": 101.95021834061136, + "grad_norm": 0.014353828504681587, + "learning_rate": 0.0006, + "loss": 4.112372398376465, + "step": 7340 + }, + { + "epoch": 101.96419213973799, + "grad_norm": 0.013797440566122532, + "learning_rate": 0.0006, + "loss": 4.2181396484375, + "step": 7341 + }, + { + "epoch": 101.97816593886463, + "grad_norm": 0.013296752236783504, + "learning_rate": 0.0006, + "loss": 4.015590190887451, + "step": 7342 + }, + { + "epoch": 101.99213973799127, + "grad_norm": 0.014963190071284771, + "learning_rate": 0.0006, + "loss": 3.9683375358581543, + "step": 7343 + }, + { + "epoch": 102.0, + "grad_norm": 0.016375407576560974, + "learning_rate": 0.0006, + "loss": 4.058782577514648, + "step": 7344 + }, + { + "epoch": 102.0, + "eval_loss": 4.608489036560059, + "eval_runtime": 57.1021, + "eval_samples_per_second": 42.765, + "eval_steps_per_second": 1.348, + "step": 7344 + }, + { + "epoch": 102.01397379912663, + "grad_norm": 0.015222231857478619, + "learning_rate": 0.0006, + "loss": 3.9541208744049072, + "step": 7345 + }, + { + "epoch": 102.02794759825328, + "grad_norm": 0.015121257863938808, + "learning_rate": 0.0006, + "loss": 4.031475067138672, + "step": 7346 + }, + { + "epoch": 102.04192139737991, + "grad_norm": 0.015098890289664268, + "learning_rate": 0.0006, + "loss": 4.025700569152832, + "step": 7347 + }, + { + "epoch": 102.05589519650655, + "grad_norm": 0.016747184097766876, + "learning_rate": 0.0006, + "loss": 4.026192665100098, + "step": 7348 + }, + { + "epoch": 102.06986899563319, + "grad_norm": 0.015377930365502834, + "learning_rate": 0.0006, + "loss": 4.099573612213135, + "step": 7349 + }, + { + "epoch": 102.08384279475983, + "grad_norm": 0.014129774644970894, + "learning_rate": 0.0006, + "loss": 3.9992661476135254, + "step": 7350 + }, + { + "epoch": 102.09781659388646, + "grad_norm": 0.016497403383255005, + "learning_rate": 0.0006, + "loss": 4.038721561431885, + "step": 7351 + }, + { + "epoch": 102.1117903930131, + "grad_norm": 0.01801230199635029, + "learning_rate": 0.0006, + "loss": 4.024539947509766, + "step": 7352 + }, + { + "epoch": 102.12576419213974, + "grad_norm": 0.016215801239013672, + "learning_rate": 0.0006, + "loss": 3.990530490875244, + "step": 7353 + }, + { + "epoch": 102.13973799126637, + "grad_norm": 0.015095847658813, + "learning_rate": 0.0006, + "loss": 4.039212226867676, + "step": 7354 + }, + { + "epoch": 102.15371179039302, + "grad_norm": 0.014441692270338535, + "learning_rate": 0.0006, + "loss": 4.033042907714844, + "step": 7355 + }, + { + "epoch": 102.16768558951965, + "grad_norm": 0.015459827147424221, + "learning_rate": 0.0006, + "loss": 3.965804100036621, + "step": 7356 + }, + { + "epoch": 102.18165938864628, + "grad_norm": 0.01581769622862339, + "learning_rate": 0.0006, + "loss": 3.9772019386291504, + "step": 7357 + }, + { + "epoch": 102.19563318777293, + "grad_norm": 0.014575488865375519, + "learning_rate": 0.0006, + "loss": 4.005141735076904, + "step": 7358 + }, + { + "epoch": 102.20960698689956, + "grad_norm": 0.01389391254633665, + "learning_rate": 0.0006, + "loss": 3.950014114379883, + "step": 7359 + }, + { + "epoch": 102.2235807860262, + "grad_norm": 0.013461007736623287, + "learning_rate": 0.0006, + "loss": 4.047972679138184, + "step": 7360 + }, + { + "epoch": 102.23755458515284, + "grad_norm": 0.015122898854315281, + "learning_rate": 0.0006, + "loss": 4.06884241104126, + "step": 7361 + }, + { + "epoch": 102.25152838427948, + "grad_norm": 0.018046488985419273, + "learning_rate": 0.0006, + "loss": 3.831460952758789, + "step": 7362 + }, + { + "epoch": 102.26550218340611, + "grad_norm": 0.01921425573527813, + "learning_rate": 0.0006, + "loss": 3.982016086578369, + "step": 7363 + }, + { + "epoch": 102.27947598253274, + "grad_norm": 0.021769993007183075, + "learning_rate": 0.0006, + "loss": 4.113831520080566, + "step": 7364 + }, + { + "epoch": 102.29344978165939, + "grad_norm": 0.02255893312394619, + "learning_rate": 0.0006, + "loss": 3.9648423194885254, + "step": 7365 + }, + { + "epoch": 102.30742358078602, + "grad_norm": 0.01851275935769081, + "learning_rate": 0.0006, + "loss": 3.963744640350342, + "step": 7366 + }, + { + "epoch": 102.32139737991267, + "grad_norm": 0.01630641333758831, + "learning_rate": 0.0006, + "loss": 3.893911600112915, + "step": 7367 + }, + { + "epoch": 102.3353711790393, + "grad_norm": 0.01931816339492798, + "learning_rate": 0.0006, + "loss": 3.9658846855163574, + "step": 7368 + }, + { + "epoch": 102.34934497816595, + "grad_norm": 0.019413312897086143, + "learning_rate": 0.0006, + "loss": 4.00821590423584, + "step": 7369 + }, + { + "epoch": 102.36331877729258, + "grad_norm": 0.01617329567670822, + "learning_rate": 0.0006, + "loss": 4.049162864685059, + "step": 7370 + }, + { + "epoch": 102.37729257641921, + "grad_norm": 0.015705779194831848, + "learning_rate": 0.0006, + "loss": 3.9896233081817627, + "step": 7371 + }, + { + "epoch": 102.39126637554585, + "grad_norm": 0.015586525201797485, + "learning_rate": 0.0006, + "loss": 3.974888801574707, + "step": 7372 + }, + { + "epoch": 102.40524017467249, + "grad_norm": 0.015439536422491074, + "learning_rate": 0.0006, + "loss": 3.9923813343048096, + "step": 7373 + }, + { + "epoch": 102.41921397379913, + "grad_norm": 0.015753930434584618, + "learning_rate": 0.0006, + "loss": 3.986898899078369, + "step": 7374 + }, + { + "epoch": 102.43318777292576, + "grad_norm": 0.014091573655605316, + "learning_rate": 0.0006, + "loss": 4.128416061401367, + "step": 7375 + }, + { + "epoch": 102.44716157205241, + "grad_norm": 0.01459917426109314, + "learning_rate": 0.0006, + "loss": 3.974200487136841, + "step": 7376 + }, + { + "epoch": 102.46113537117904, + "grad_norm": 0.014099135994911194, + "learning_rate": 0.0006, + "loss": 4.034481525421143, + "step": 7377 + }, + { + "epoch": 102.47510917030567, + "grad_norm": 0.013483141548931599, + "learning_rate": 0.0006, + "loss": 3.901756763458252, + "step": 7378 + }, + { + "epoch": 102.48908296943232, + "grad_norm": 0.013499923050403595, + "learning_rate": 0.0006, + "loss": 3.965855598449707, + "step": 7379 + }, + { + "epoch": 102.50305676855895, + "grad_norm": 0.01356455497443676, + "learning_rate": 0.0006, + "loss": 3.8891913890838623, + "step": 7380 + }, + { + "epoch": 102.5170305676856, + "grad_norm": 0.013884859159588814, + "learning_rate": 0.0006, + "loss": 4.171500205993652, + "step": 7381 + }, + { + "epoch": 102.53100436681223, + "grad_norm": 0.014240458607673645, + "learning_rate": 0.0006, + "loss": 4.021063804626465, + "step": 7382 + }, + { + "epoch": 102.54497816593886, + "grad_norm": 0.015500220470130444, + "learning_rate": 0.0006, + "loss": 4.073713302612305, + "step": 7383 + }, + { + "epoch": 102.5589519650655, + "grad_norm": 0.01415219809859991, + "learning_rate": 0.0006, + "loss": 4.095281600952148, + "step": 7384 + }, + { + "epoch": 102.57292576419214, + "grad_norm": 0.01237252727150917, + "learning_rate": 0.0006, + "loss": 3.929661750793457, + "step": 7385 + }, + { + "epoch": 102.58689956331878, + "grad_norm": 0.013517644256353378, + "learning_rate": 0.0006, + "loss": 4.0342583656311035, + "step": 7386 + }, + { + "epoch": 102.60087336244541, + "grad_norm": 0.013656767085194588, + "learning_rate": 0.0006, + "loss": 4.05634880065918, + "step": 7387 + }, + { + "epoch": 102.61484716157206, + "grad_norm": 0.01358469296246767, + "learning_rate": 0.0006, + "loss": 4.069542407989502, + "step": 7388 + }, + { + "epoch": 102.62882096069869, + "grad_norm": 0.013493471778929234, + "learning_rate": 0.0006, + "loss": 3.9387717247009277, + "step": 7389 + }, + { + "epoch": 102.64279475982532, + "grad_norm": 0.015176896937191486, + "learning_rate": 0.0006, + "loss": 3.9376473426818848, + "step": 7390 + }, + { + "epoch": 102.65676855895197, + "grad_norm": 0.013986237347126007, + "learning_rate": 0.0006, + "loss": 3.96488618850708, + "step": 7391 + }, + { + "epoch": 102.6707423580786, + "grad_norm": 0.013385794125497341, + "learning_rate": 0.0006, + "loss": 4.048396110534668, + "step": 7392 + }, + { + "epoch": 102.68471615720524, + "grad_norm": 0.012358390726149082, + "learning_rate": 0.0006, + "loss": 3.950759172439575, + "step": 7393 + }, + { + "epoch": 102.69868995633188, + "grad_norm": 0.01348735112696886, + "learning_rate": 0.0006, + "loss": 4.1605658531188965, + "step": 7394 + }, + { + "epoch": 102.71266375545852, + "grad_norm": 0.01473618671298027, + "learning_rate": 0.0006, + "loss": 4.0965046882629395, + "step": 7395 + }, + { + "epoch": 102.72663755458515, + "grad_norm": 0.014261185191571712, + "learning_rate": 0.0006, + "loss": 4.105083465576172, + "step": 7396 + }, + { + "epoch": 102.74061135371178, + "grad_norm": 0.013883501291275024, + "learning_rate": 0.0006, + "loss": 4.028477668762207, + "step": 7397 + }, + { + "epoch": 102.75458515283843, + "grad_norm": 0.0138440802693367, + "learning_rate": 0.0006, + "loss": 4.05599308013916, + "step": 7398 + }, + { + "epoch": 102.76855895196506, + "grad_norm": 0.014218976721167564, + "learning_rate": 0.0006, + "loss": 4.154707908630371, + "step": 7399 + }, + { + "epoch": 102.78253275109171, + "grad_norm": 0.013328758999705315, + "learning_rate": 0.0006, + "loss": 4.060786247253418, + "step": 7400 + }, + { + "epoch": 102.79650655021834, + "grad_norm": 0.013351088389754295, + "learning_rate": 0.0006, + "loss": 4.045477867126465, + "step": 7401 + }, + { + "epoch": 102.81048034934499, + "grad_norm": 0.014380485750734806, + "learning_rate": 0.0006, + "loss": 4.032599449157715, + "step": 7402 + }, + { + "epoch": 102.82445414847162, + "grad_norm": 0.014046687632799149, + "learning_rate": 0.0006, + "loss": 4.006680488586426, + "step": 7403 + }, + { + "epoch": 102.83842794759825, + "grad_norm": 0.013186605647206306, + "learning_rate": 0.0006, + "loss": 4.0323991775512695, + "step": 7404 + }, + { + "epoch": 102.8524017467249, + "grad_norm": 0.013620814308524132, + "learning_rate": 0.0006, + "loss": 3.9950883388519287, + "step": 7405 + }, + { + "epoch": 102.86637554585153, + "grad_norm": 0.013619142584502697, + "learning_rate": 0.0006, + "loss": 4.052000999450684, + "step": 7406 + }, + { + "epoch": 102.88034934497817, + "grad_norm": 0.012630553916096687, + "learning_rate": 0.0006, + "loss": 4.104743957519531, + "step": 7407 + }, + { + "epoch": 102.8943231441048, + "grad_norm": 0.013503247871994972, + "learning_rate": 0.0006, + "loss": 4.055751800537109, + "step": 7408 + }, + { + "epoch": 102.90829694323143, + "grad_norm": 0.01509852521121502, + "learning_rate": 0.0006, + "loss": 4.064550399780273, + "step": 7409 + }, + { + "epoch": 102.92227074235808, + "grad_norm": 0.015878785401582718, + "learning_rate": 0.0006, + "loss": 4.0609917640686035, + "step": 7410 + }, + { + "epoch": 102.93624454148471, + "grad_norm": 0.015079273842275143, + "learning_rate": 0.0006, + "loss": 3.9825186729431152, + "step": 7411 + }, + { + "epoch": 102.95021834061136, + "grad_norm": 0.014393322169780731, + "learning_rate": 0.0006, + "loss": 4.089018821716309, + "step": 7412 + }, + { + "epoch": 102.96419213973799, + "grad_norm": 0.01369137316942215, + "learning_rate": 0.0006, + "loss": 4.038545608520508, + "step": 7413 + }, + { + "epoch": 102.97816593886463, + "grad_norm": 0.013304094783961773, + "learning_rate": 0.0006, + "loss": 4.075267314910889, + "step": 7414 + }, + { + "epoch": 102.99213973799127, + "grad_norm": 0.01388571783900261, + "learning_rate": 0.0006, + "loss": 4.054773330688477, + "step": 7415 + }, + { + "epoch": 103.0, + "grad_norm": 0.015209192410111427, + "learning_rate": 0.0006, + "loss": 3.9846372604370117, + "step": 7416 + }, + { + "epoch": 103.0, + "eval_loss": 4.649291515350342, + "eval_runtime": 56.474, + "eval_samples_per_second": 43.241, + "eval_steps_per_second": 1.363, + "step": 7416 + }, + { + "epoch": 103.01397379912663, + "grad_norm": 0.014897244982421398, + "learning_rate": 0.0006, + "loss": 3.905099868774414, + "step": 7417 + }, + { + "epoch": 103.02794759825328, + "grad_norm": 0.015420001931488514, + "learning_rate": 0.0006, + "loss": 3.985867500305176, + "step": 7418 + }, + { + "epoch": 103.04192139737991, + "grad_norm": 0.016167795285582542, + "learning_rate": 0.0006, + "loss": 3.8834733963012695, + "step": 7419 + }, + { + "epoch": 103.05589519650655, + "grad_norm": 0.017051100730895996, + "learning_rate": 0.0006, + "loss": 4.044392108917236, + "step": 7420 + }, + { + "epoch": 103.06986899563319, + "grad_norm": 0.016036640852689743, + "learning_rate": 0.0006, + "loss": 4.0110883712768555, + "step": 7421 + }, + { + "epoch": 103.08384279475983, + "grad_norm": 0.016148710623383522, + "learning_rate": 0.0006, + "loss": 4.0064544677734375, + "step": 7422 + }, + { + "epoch": 103.09781659388646, + "grad_norm": 0.014666100963950157, + "learning_rate": 0.0006, + "loss": 3.9640684127807617, + "step": 7423 + }, + { + "epoch": 103.1117903930131, + "grad_norm": 0.015464609488844872, + "learning_rate": 0.0006, + "loss": 4.069262504577637, + "step": 7424 + }, + { + "epoch": 103.12576419213974, + "grad_norm": 0.01731940731406212, + "learning_rate": 0.0006, + "loss": 3.9752140045166016, + "step": 7425 + }, + { + "epoch": 103.13973799126637, + "grad_norm": 0.017115836963057518, + "learning_rate": 0.0006, + "loss": 3.9323995113372803, + "step": 7426 + }, + { + "epoch": 103.15371179039302, + "grad_norm": 0.016439078375697136, + "learning_rate": 0.0006, + "loss": 4.088135719299316, + "step": 7427 + }, + { + "epoch": 103.16768558951965, + "grad_norm": 0.015544523485004902, + "learning_rate": 0.0006, + "loss": 4.152291297912598, + "step": 7428 + }, + { + "epoch": 103.18165938864628, + "grad_norm": 0.016065534204244614, + "learning_rate": 0.0006, + "loss": 4.069518566131592, + "step": 7429 + }, + { + "epoch": 103.19563318777293, + "grad_norm": 0.015838216990232468, + "learning_rate": 0.0006, + "loss": 3.9768013954162598, + "step": 7430 + }, + { + "epoch": 103.20960698689956, + "grad_norm": 0.01694624498486519, + "learning_rate": 0.0006, + "loss": 3.974421977996826, + "step": 7431 + }, + { + "epoch": 103.2235807860262, + "grad_norm": 0.017309164628386497, + "learning_rate": 0.0006, + "loss": 4.101884841918945, + "step": 7432 + }, + { + "epoch": 103.23755458515284, + "grad_norm": 0.017330938950181007, + "learning_rate": 0.0006, + "loss": 4.018990516662598, + "step": 7433 + }, + { + "epoch": 103.25152838427948, + "grad_norm": 0.01596875861287117, + "learning_rate": 0.0006, + "loss": 4.1121063232421875, + "step": 7434 + }, + { + "epoch": 103.26550218340611, + "grad_norm": 0.017698034644126892, + "learning_rate": 0.0006, + "loss": 4.090383052825928, + "step": 7435 + }, + { + "epoch": 103.27947598253274, + "grad_norm": 0.018080011010169983, + "learning_rate": 0.0006, + "loss": 3.980351448059082, + "step": 7436 + }, + { + "epoch": 103.29344978165939, + "grad_norm": 0.01784810796380043, + "learning_rate": 0.0006, + "loss": 3.836942672729492, + "step": 7437 + }, + { + "epoch": 103.30742358078602, + "grad_norm": 0.01571204699575901, + "learning_rate": 0.0006, + "loss": 3.9809398651123047, + "step": 7438 + }, + { + "epoch": 103.32139737991267, + "grad_norm": 0.015144702047109604, + "learning_rate": 0.0006, + "loss": 4.122274875640869, + "step": 7439 + }, + { + "epoch": 103.3353711790393, + "grad_norm": 0.017243649810552597, + "learning_rate": 0.0006, + "loss": 3.9569320678710938, + "step": 7440 + }, + { + "epoch": 103.34934497816595, + "grad_norm": 0.017546052113175392, + "learning_rate": 0.0006, + "loss": 3.8998186588287354, + "step": 7441 + }, + { + "epoch": 103.36331877729258, + "grad_norm": 0.016121823340654373, + "learning_rate": 0.0006, + "loss": 4.154199600219727, + "step": 7442 + }, + { + "epoch": 103.37729257641921, + "grad_norm": 0.01464556809514761, + "learning_rate": 0.0006, + "loss": 3.8896102905273438, + "step": 7443 + }, + { + "epoch": 103.39126637554585, + "grad_norm": 0.015656189993023872, + "learning_rate": 0.0006, + "loss": 3.9351630210876465, + "step": 7444 + }, + { + "epoch": 103.40524017467249, + "grad_norm": 0.01604314334690571, + "learning_rate": 0.0006, + "loss": 4.0186238288879395, + "step": 7445 + }, + { + "epoch": 103.41921397379913, + "grad_norm": 0.014529184438288212, + "learning_rate": 0.0006, + "loss": 4.003365516662598, + "step": 7446 + }, + { + "epoch": 103.43318777292576, + "grad_norm": 0.015263797715306282, + "learning_rate": 0.0006, + "loss": 4.097224712371826, + "step": 7447 + }, + { + "epoch": 103.44716157205241, + "grad_norm": 0.015169554390013218, + "learning_rate": 0.0006, + "loss": 4.023651123046875, + "step": 7448 + }, + { + "epoch": 103.46113537117904, + "grad_norm": 0.014693229459226131, + "learning_rate": 0.0006, + "loss": 4.095487117767334, + "step": 7449 + }, + { + "epoch": 103.47510917030567, + "grad_norm": 0.015125681646168232, + "learning_rate": 0.0006, + "loss": 3.9849324226379395, + "step": 7450 + }, + { + "epoch": 103.48908296943232, + "grad_norm": 0.014218835160136223, + "learning_rate": 0.0006, + "loss": 4.104394912719727, + "step": 7451 + }, + { + "epoch": 103.50305676855895, + "grad_norm": 0.014997185207903385, + "learning_rate": 0.0006, + "loss": 3.9534459114074707, + "step": 7452 + }, + { + "epoch": 103.5170305676856, + "grad_norm": 0.014991429634392262, + "learning_rate": 0.0006, + "loss": 4.198554992675781, + "step": 7453 + }, + { + "epoch": 103.53100436681223, + "grad_norm": 0.013290980830788612, + "learning_rate": 0.0006, + "loss": 3.857856273651123, + "step": 7454 + }, + { + "epoch": 103.54497816593886, + "grad_norm": 0.013200514949858189, + "learning_rate": 0.0006, + "loss": 3.96994948387146, + "step": 7455 + }, + { + "epoch": 103.5589519650655, + "grad_norm": 0.01466651912778616, + "learning_rate": 0.0006, + "loss": 4.026296138763428, + "step": 7456 + }, + { + "epoch": 103.57292576419214, + "grad_norm": 0.015408058650791645, + "learning_rate": 0.0006, + "loss": 3.9223692417144775, + "step": 7457 + }, + { + "epoch": 103.58689956331878, + "grad_norm": 0.014617101289331913, + "learning_rate": 0.0006, + "loss": 3.9703640937805176, + "step": 7458 + }, + { + "epoch": 103.60087336244541, + "grad_norm": 0.013879700563848019, + "learning_rate": 0.0006, + "loss": 3.9978384971618652, + "step": 7459 + }, + { + "epoch": 103.61484716157206, + "grad_norm": 0.0155031718313694, + "learning_rate": 0.0006, + "loss": 4.004831314086914, + "step": 7460 + }, + { + "epoch": 103.62882096069869, + "grad_norm": 0.01537798810750246, + "learning_rate": 0.0006, + "loss": 4.076343059539795, + "step": 7461 + }, + { + "epoch": 103.64279475982532, + "grad_norm": 0.012790355831384659, + "learning_rate": 0.0006, + "loss": 4.096382141113281, + "step": 7462 + }, + { + "epoch": 103.65676855895197, + "grad_norm": 0.012812363915145397, + "learning_rate": 0.0006, + "loss": 4.065099716186523, + "step": 7463 + }, + { + "epoch": 103.6707423580786, + "grad_norm": 0.01436334103345871, + "learning_rate": 0.0006, + "loss": 3.9989516735076904, + "step": 7464 + }, + { + "epoch": 103.68471615720524, + "grad_norm": 0.014180350117385387, + "learning_rate": 0.0006, + "loss": 4.0170698165893555, + "step": 7465 + }, + { + "epoch": 103.69868995633188, + "grad_norm": 0.01401323638856411, + "learning_rate": 0.0006, + "loss": 4.02116584777832, + "step": 7466 + }, + { + "epoch": 103.71266375545852, + "grad_norm": 0.01357064489275217, + "learning_rate": 0.0006, + "loss": 4.083914279937744, + "step": 7467 + }, + { + "epoch": 103.72663755458515, + "grad_norm": 0.013478172942996025, + "learning_rate": 0.0006, + "loss": 3.946953773498535, + "step": 7468 + }, + { + "epoch": 103.74061135371178, + "grad_norm": 0.014460090547800064, + "learning_rate": 0.0006, + "loss": 3.8642897605895996, + "step": 7469 + }, + { + "epoch": 103.75458515283843, + "grad_norm": 0.015068473294377327, + "learning_rate": 0.0006, + "loss": 4.096778869628906, + "step": 7470 + }, + { + "epoch": 103.76855895196506, + "grad_norm": 0.01593821309506893, + "learning_rate": 0.0006, + "loss": 3.9700300693511963, + "step": 7471 + }, + { + "epoch": 103.78253275109171, + "grad_norm": 0.016736770048737526, + "learning_rate": 0.0006, + "loss": 4.045576572418213, + "step": 7472 + }, + { + "epoch": 103.79650655021834, + "grad_norm": 0.01562635414302349, + "learning_rate": 0.0006, + "loss": 4.108773231506348, + "step": 7473 + }, + { + "epoch": 103.81048034934499, + "grad_norm": 0.014363548718392849, + "learning_rate": 0.0006, + "loss": 4.199751853942871, + "step": 7474 + }, + { + "epoch": 103.82445414847162, + "grad_norm": 0.013952836394309998, + "learning_rate": 0.0006, + "loss": 4.087684631347656, + "step": 7475 + }, + { + "epoch": 103.83842794759825, + "grad_norm": 0.015203139744699001, + "learning_rate": 0.0006, + "loss": 4.094961166381836, + "step": 7476 + }, + { + "epoch": 103.8524017467249, + "grad_norm": 0.015558234415948391, + "learning_rate": 0.0006, + "loss": 4.041770935058594, + "step": 7477 + }, + { + "epoch": 103.86637554585153, + "grad_norm": 0.01485239528119564, + "learning_rate": 0.0006, + "loss": 3.9825830459594727, + "step": 7478 + }, + { + "epoch": 103.88034934497817, + "grad_norm": 0.013139205053448677, + "learning_rate": 0.0006, + "loss": 4.084344863891602, + "step": 7479 + }, + { + "epoch": 103.8943231441048, + "grad_norm": 0.013366639614105225, + "learning_rate": 0.0006, + "loss": 4.10720682144165, + "step": 7480 + }, + { + "epoch": 103.90829694323143, + "grad_norm": 0.012898124754428864, + "learning_rate": 0.0006, + "loss": 4.066727161407471, + "step": 7481 + }, + { + "epoch": 103.92227074235808, + "grad_norm": 0.012501404620707035, + "learning_rate": 0.0006, + "loss": 4.064421653747559, + "step": 7482 + }, + { + "epoch": 103.93624454148471, + "grad_norm": 0.013129777275025845, + "learning_rate": 0.0006, + "loss": 3.908412456512451, + "step": 7483 + }, + { + "epoch": 103.95021834061136, + "grad_norm": 0.01346709206700325, + "learning_rate": 0.0006, + "loss": 4.070006847381592, + "step": 7484 + }, + { + "epoch": 103.96419213973799, + "grad_norm": 0.013403614051640034, + "learning_rate": 0.0006, + "loss": 3.978532314300537, + "step": 7485 + }, + { + "epoch": 103.97816593886463, + "grad_norm": 0.013506206683814526, + "learning_rate": 0.0006, + "loss": 4.025320053100586, + "step": 7486 + }, + { + "epoch": 103.99213973799127, + "grad_norm": 0.015543299727141857, + "learning_rate": 0.0006, + "loss": 4.066288948059082, + "step": 7487 + }, + { + "epoch": 104.0, + "grad_norm": 0.017104074358940125, + "learning_rate": 0.0006, + "loss": 3.9784560203552246, + "step": 7488 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.192397895785513e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-7488/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-7488/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-7488/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja b/runs/l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/config.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/config.json new file mode 100644 index 0000000000000000000000000000000000000000..22f250953c6ce8fc91ee4411856dd73041a415e2 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151673 +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/generation_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/model.safetensors b/runs/l2r50-i2-fulle-lm/checkpoint-864/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54b5c83669af15f0b9140b26ae8ae51ec68dca54 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:179bf5635dd10044802d1dc44ed24b19194380f3efc90306146f615333929f19 +size 583360328 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt b/runs/l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d1f8285526f273614069da6f6ec8b57e2be1aa31 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e1c6b2105f2225f0b40ba61fdf70d02a365d80a1338b8755155116ec1401037 +size 1166833530 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..2850db5a8b3c23d091b3fd3a0338fa0da162e9cb --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d88b5132c373442240592255aff89e931e66ca54a5309ba0106d24acc1c940e8 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..01e3351fc2ba528e4ec4ec330aa331839cf9147b --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f969ec38f9da2e2770ade6f42ea9a45cc7aea3a552040d097648e88ad7651bc0 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..e3400b288bec9024746b50e3d0efc784072fe544 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b28227375ca73bdf9d25dbde42e64d4ea3bea06c551f80c1045cbb59913998a9 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..0784197f679ee63333ba951dc2ea04a15f9a85df --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89301a7e872e726f45ed2ce4fe0342d3a8e335a1ba8b8f8575869beec7fb5f36 +size 15024 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt b/runs/l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f21fc027a37dccb66b77a4752de227c557d8ff04 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1a1dda88112968a1342ac518f362b4f5e138437f0b670c818ad8493b91ca49d +size 1064 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..53786016f070e11184813f3c83101dc5e93e392e --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3c63fda1646a555448e8f03662542e04b11d103721319e4d9b267d7da03ceb8 +size 11424463 diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..98e2fb9b2cc54785dc9d871b8e503f8b8f2defe7 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json @@ -0,0 +1,20 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json b/runs/l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..31ae1d5dacebb36361cb567c4f9d884367b89410 --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json @@ -0,0 +1,6170 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 12.0, + "eval_steps": 500, + "global_step": 864, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.1739506870508194, + "learning_rate": 0.0, + "loss": 11.99299430847168, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.17216597497463226, + "learning_rate": 5.999999999999999e-06, + "loss": 11.988704681396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.17459715902805328, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.945581436157227, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.17577414214611053, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.859772682189941, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.16163651645183563, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.744415283203125, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.14409932494163513, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.618099212646484, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.13079848885536194, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.505935668945312, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.11964613944292068, + "learning_rate": 4.2e-05, + "loss": 11.416532516479492, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.11345142871141434, + "learning_rate": 4.7999999999999994e-05, + "loss": 11.33206558227539, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.11033952981233597, + "learning_rate": 5.399999999999999e-05, + "loss": 11.259052276611328, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10776489973068237, + "learning_rate": 5.9999999999999995e-05, + "loss": 11.203384399414062, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.10721946507692337, + "learning_rate": 6.599999999999999e-05, + "loss": 11.153295516967773, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10673430562019348, + "learning_rate": 7.199999999999999e-05, + "loss": 11.110200881958008, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10630254447460175, + "learning_rate": 7.8e-05, + "loss": 11.067477226257324, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10631615668535233, + "learning_rate": 8.4e-05, + "loss": 11.022941589355469, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10609443485736847, + "learning_rate": 8.999999999999999e-05, + "loss": 10.970746040344238, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.10648232698440552, + "learning_rate": 9.599999999999999e-05, + "loss": 10.91267204284668, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.10584589838981628, + "learning_rate": 0.000102, + "loss": 10.854547500610352, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.10591720044612885, + "learning_rate": 0.00010799999999999998, + "loss": 10.788917541503906, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.10611672699451447, + "learning_rate": 0.00011399999999999999, + "loss": 10.717556953430176, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.10455359518527985, + "learning_rate": 0.00011999999999999999, + "loss": 10.650186538696289, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.1047525629401207, + "learning_rate": 0.00012599999999999997, + "loss": 10.56808090209961, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.10511605441570282, + "learning_rate": 0.00013199999999999998, + "loss": 10.484028816223145, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.10467177629470825, + "learning_rate": 0.000138, + "loss": 10.399947166442871, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.10432136058807373, + "learning_rate": 0.00014399999999999998, + "loss": 10.312714576721191, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.10527262091636658, + "learning_rate": 0.00015, + "loss": 10.212261199951172, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.103610560297966, + "learning_rate": 0.000156, + "loss": 10.132034301757812, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.10553059726953506, + "learning_rate": 0.000162, + "loss": 10.020434379577637, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.1037299782037735, + "learning_rate": 0.000168, + "loss": 9.934466361999512, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.1042739674448967, + "learning_rate": 0.00017399999999999997, + "loss": 9.822539329528809, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.10300865024328232, + "learning_rate": 0.00017999999999999998, + "loss": 9.734994888305664, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.10291654616594315, + "learning_rate": 0.000186, + "loss": 9.624505996704102, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.10210851579904556, + "learning_rate": 0.00019199999999999998, + "loss": 9.518205642700195, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.10222529619932175, + "learning_rate": 0.000198, + "loss": 9.412222862243652, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.09992019087076187, + "learning_rate": 0.000204, + "loss": 9.326566696166992, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.09931084513664246, + "learning_rate": 0.00020999999999999998, + "loss": 9.210111618041992, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.09928256273269653, + "learning_rate": 0.00021599999999999996, + "loss": 9.100493431091309, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.09843816608190536, + "learning_rate": 0.00022199999999999998, + "loss": 8.996543884277344, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.09825374186038971, + "learning_rate": 0.00022799999999999999, + "loss": 8.888991355895996, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.09602385759353638, + "learning_rate": 0.000234, + "loss": 8.802907943725586, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.09414064139127731, + "learning_rate": 0.00023999999999999998, + "loss": 8.715423583984375, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.09273893386125565, + "learning_rate": 0.00024599999999999996, + "loss": 8.61374282836914, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.09079599380493164, + "learning_rate": 0.00025199999999999995, + "loss": 8.502904891967773, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.08605825901031494, + "learning_rate": 0.000258, + "loss": 8.453008651733398, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.08338996022939682, + "learning_rate": 0.00026399999999999997, + "loss": 8.365163803100586, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.08150867372751236, + "learning_rate": 0.00027, + "loss": 8.285372734069824, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.07794807851314545, + "learning_rate": 0.000276, + "loss": 8.185811996459961, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.07592720538377762, + "learning_rate": 0.00028199999999999997, + "loss": 8.098482131958008, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0709897130727768, + "learning_rate": 0.00028799999999999995, + "loss": 8.042091369628906, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.06582305580377579, + "learning_rate": 0.000294, + "loss": 7.972844123840332, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.059452351182699203, + "learning_rate": 0.0003, + "loss": 7.924579620361328, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.052426718175411224, + "learning_rate": 0.00030599999999999996, + "loss": 7.876336097717285, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.04768150672316551, + "learning_rate": 0.000312, + "loss": 7.8102707862854, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.03970075398683548, + "learning_rate": 0.000318, + "loss": 7.791913986206055, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.03559752553701401, + "learning_rate": 0.000324, + "loss": 7.756960868835449, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.029321353882551193, + "learning_rate": 0.00033, + "loss": 7.724878787994385, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.023485427722334862, + "learning_rate": 0.000336, + "loss": 7.72553014755249, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.019319474697113037, + "learning_rate": 0.00034199999999999996, + "loss": 7.709493160247803, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.016705691814422607, + "learning_rate": 0.00034799999999999995, + "loss": 7.6873297691345215, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.017667734995484352, + "learning_rate": 0.00035399999999999993, + "loss": 7.677059173583984, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.01671992614865303, + "learning_rate": 0.00035999999999999997, + "loss": 7.643512725830078, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.012288172729313374, + "learning_rate": 0.00036599999999999995, + "loss": 7.658343315124512, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.012004776857793331, + "learning_rate": 0.000372, + "loss": 7.649759292602539, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.012487445957958698, + "learning_rate": 0.00037799999999999997, + "loss": 7.631503105163574, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.01489241048693657, + "learning_rate": 0.00038399999999999996, + "loss": 7.647110462188721, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01428980939090252, + "learning_rate": 0.00039, + "loss": 7.636983871459961, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.010711144655942917, + "learning_rate": 0.000396, + "loss": 7.621130466461182, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.010178886353969574, + "learning_rate": 0.000402, + "loss": 7.593924522399902, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.010604874230921268, + "learning_rate": 0.000408, + "loss": 7.6015496253967285, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.009751426056027412, + "learning_rate": 0.0004139999999999999, + "loss": 7.571964263916016, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.009816481731832027, + "learning_rate": 0.00041999999999999996, + "loss": 7.57344913482666, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.008665637113153934, + "learning_rate": 0.00042599999999999995, + "loss": 7.579050064086914, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.581204891204834, + "eval_runtime": 59.0091, + "eval_samples_per_second": 41.383, + "eval_steps_per_second": 1.305, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.009684201329946518, + "learning_rate": 0.00043199999999999993, + "loss": 7.568594932556152, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.025783710181713104, + "learning_rate": 0.00043799999999999997, + "loss": 7.578948497772217, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.0327819399535656, + "learning_rate": 0.00044399999999999995, + "loss": 7.548173904418945, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.021604498848319054, + "learning_rate": 0.00045, + "loss": 7.552953720092773, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.010230105370283127, + "learning_rate": 0.00045599999999999997, + "loss": 7.5213212966918945, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.023315589874982834, + "learning_rate": 0.00046199999999999995, + "loss": 7.501640319824219, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.01109944935888052, + "learning_rate": 0.000468, + "loss": 7.539519786834717, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.026093758642673492, + "learning_rate": 0.000474, + "loss": 7.525928974151611, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.03274895250797272, + "learning_rate": 0.00047999999999999996, + "loss": 7.535602569580078, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01543602254241705, + "learning_rate": 0.000486, + "loss": 7.513683795928955, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.028377249836921692, + "learning_rate": 0.0004919999999999999, + "loss": 7.49384880065918, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.04121481627225876, + "learning_rate": 0.000498, + "loss": 7.488447189331055, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.013282302767038345, + "learning_rate": 0.0005039999999999999, + "loss": 7.463010787963867, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.03200364485383034, + "learning_rate": 0.0005099999999999999, + "loss": 7.473515033721924, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.020768029615283012, + "learning_rate": 0.000516, + "loss": 7.464688777923584, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.021715788170695305, + "learning_rate": 0.000522, + "loss": 7.4835028648376465, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.026255754753947258, + "learning_rate": 0.0005279999999999999, + "loss": 7.443392753601074, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.022325951606035233, + "learning_rate": 0.000534, + "loss": 7.457373142242432, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009270387701690197, + "learning_rate": 0.00054, + "loss": 7.455986976623535, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.02369544468820095, + "learning_rate": 0.0005459999999999999, + "loss": 7.442216873168945, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.008474506437778473, + "learning_rate": 0.000552, + "loss": 7.4294915199279785, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.017043249681591988, + "learning_rate": 0.000558, + "loss": 7.406932830810547, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.011813412420451641, + "learning_rate": 0.0005639999999999999, + "loss": 7.415900230407715, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.01086883433163166, + "learning_rate": 0.00057, + "loss": 7.452998638153076, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.014700951054692268, + "learning_rate": 0.0005759999999999999, + "loss": 7.410249710083008, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.014512976631522179, + "learning_rate": 0.0005819999999999999, + "loss": 7.429132461547852, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.012519920244812965, + "learning_rate": 0.000588, + "loss": 7.430817604064941, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007132664322853088, + "learning_rate": 0.0005939999999999999, + "loss": 7.405234336853027, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.019489388912916183, + "learning_rate": 0.0006, + "loss": 7.427791595458984, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.027252664789557457, + "learning_rate": 0.0006, + "loss": 7.41761589050293, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.02501637488603592, + "learning_rate": 0.0006, + "loss": 7.416735649108887, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.00989772193133831, + "learning_rate": 0.0006, + "loss": 7.3756632804870605, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.01622949168086052, + "learning_rate": 0.0006, + "loss": 7.400811672210693, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.012407763861119747, + "learning_rate": 0.0006, + "loss": 7.37675666809082, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.010353565216064453, + "learning_rate": 0.0006, + "loss": 7.4013142585754395, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0120553532615304, + "learning_rate": 0.0006, + "loss": 7.379873275756836, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01610225811600685, + "learning_rate": 0.0006, + "loss": 7.371631145477295, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.01159879844635725, + "learning_rate": 0.0006, + "loss": 7.389007568359375, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.008013816550374031, + "learning_rate": 0.0006, + "loss": 7.395920753479004, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.016334809362888336, + "learning_rate": 0.0006, + "loss": 7.393121719360352, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.018821721896529198, + "learning_rate": 0.0006, + "loss": 7.367543697357178, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.014088994823396206, + "learning_rate": 0.0006, + "loss": 7.385537624359131, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.005864050704985857, + "learning_rate": 0.0006, + "loss": 7.365346908569336, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.011745430529117584, + "learning_rate": 0.0006, + "loss": 7.390227317810059, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.017205573618412018, + "learning_rate": 0.0006, + "loss": 7.3558549880981445, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021256346255540848, + "learning_rate": 0.0006, + "loss": 7.377346992492676, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.019335204735398293, + "learning_rate": 0.0006, + "loss": 7.346194267272949, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.008134003728628159, + "learning_rate": 0.0006, + "loss": 7.3063578605651855, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.017613831907510757, + "learning_rate": 0.0006, + "loss": 7.338171005249023, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.00890734139829874, + "learning_rate": 0.0006, + "loss": 7.355384826660156, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013448155485093594, + "learning_rate": 0.0006, + "loss": 7.344296455383301, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.009241633117198944, + "learning_rate": 0.0006, + "loss": 7.3454060554504395, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.018561311066150665, + "learning_rate": 0.0006, + "loss": 7.32348108291626, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.025667704641819, + "learning_rate": 0.0006, + "loss": 7.356142520904541, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.036232151091098785, + "learning_rate": 0.0006, + "loss": 7.363539695739746, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.059483714401721954, + "learning_rate": 0.0006, + "loss": 7.402331352233887, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.05365009233355522, + "learning_rate": 0.0006, + "loss": 7.40457820892334, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.02738415263593197, + "learning_rate": 0.0006, + "loss": 7.3756208419799805, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.035017091780900955, + "learning_rate": 0.0006, + "loss": 7.379263877868652, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.02819863148033619, + "learning_rate": 0.0006, + "loss": 7.327673435211182, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.017754238098859787, + "learning_rate": 0.0006, + "loss": 7.333420276641846, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.024685733020305634, + "learning_rate": 0.0006, + "loss": 7.34493350982666, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.019426532089710236, + "learning_rate": 0.0006, + "loss": 7.328448295593262, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.02106391079723835, + "learning_rate": 0.0006, + "loss": 7.327917098999023, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.02056879736483097, + "learning_rate": 0.0006, + "loss": 7.3311991691589355, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.016865141689777374, + "learning_rate": 0.0006, + "loss": 7.312766075134277, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.017549855634570122, + "learning_rate": 0.0006, + "loss": 7.31191349029541, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.015930157154798508, + "learning_rate": 0.0006, + "loss": 7.330901622772217, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.013283558189868927, + "learning_rate": 0.0006, + "loss": 7.298318386077881, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.012023469433188438, + "learning_rate": 0.0006, + "loss": 7.282337188720703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.01459511648863554, + "learning_rate": 0.0006, + "loss": 7.282924175262451, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.017474235966801643, + "learning_rate": 0.0006, + "loss": 7.254333972930908, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.318708896636963, + "eval_runtime": 59.6555, + "eval_samples_per_second": 40.935, + "eval_steps_per_second": 1.291, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.019757507368922234, + "learning_rate": 0.0006, + "loss": 7.30288553237915, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.014571248553693295, + "learning_rate": 0.0006, + "loss": 7.274681568145752, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.015744198113679886, + "learning_rate": 0.0006, + "loss": 7.278632640838623, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.028048858046531677, + "learning_rate": 0.0006, + "loss": 7.26558780670166, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.06547307223081589, + "learning_rate": 0.0006, + "loss": 7.32567834854126, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.05969763547182083, + "learning_rate": 0.0006, + "loss": 7.395836353302002, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.032574281096458435, + "learning_rate": 0.0006, + "loss": 7.318150997161865, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.029452962800860405, + "learning_rate": 0.0006, + "loss": 7.308066368103027, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.03161991387605667, + "learning_rate": 0.0006, + "loss": 7.306290626525879, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.027096295729279518, + "learning_rate": 0.0006, + "loss": 7.297796249389648, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.028657056391239166, + "learning_rate": 0.0006, + "loss": 7.317424774169922, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.02046995982527733, + "learning_rate": 0.0006, + "loss": 7.297986030578613, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.02220369316637516, + "learning_rate": 0.0006, + "loss": 7.268237590789795, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.018381357192993164, + "learning_rate": 0.0006, + "loss": 7.259415626525879, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.0204413291066885, + "learning_rate": 0.0006, + "loss": 7.258173942565918, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.015874288976192474, + "learning_rate": 0.0006, + "loss": 7.281834602355957, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.01634068600833416, + "learning_rate": 0.0006, + "loss": 7.2469868659973145, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01728747971355915, + "learning_rate": 0.0006, + "loss": 7.280995845794678, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.016341188922524452, + "learning_rate": 0.0006, + "loss": 7.273357391357422, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.016125807538628578, + "learning_rate": 0.0006, + "loss": 7.273510932922363, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01629287749528885, + "learning_rate": 0.0006, + "loss": 7.267029762268066, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.015296131372451782, + "learning_rate": 0.0006, + "loss": 7.211989402770996, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.01709570363163948, + "learning_rate": 0.0006, + "loss": 7.242851734161377, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.012552388943731785, + "learning_rate": 0.0006, + "loss": 7.241696834564209, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.014571291394531727, + "learning_rate": 0.0006, + "loss": 7.270681381225586, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.013389437459409237, + "learning_rate": 0.0006, + "loss": 7.191600799560547, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.011895314790308475, + "learning_rate": 0.0006, + "loss": 7.213610649108887, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.01345642190426588, + "learning_rate": 0.0006, + "loss": 7.193787574768066, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.0091794328764081, + "learning_rate": 0.0006, + "loss": 7.233364105224609, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.013767086900770664, + "learning_rate": 0.0006, + "loss": 7.217272758483887, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.015134960412979126, + "learning_rate": 0.0006, + "loss": 7.192791938781738, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.01790648326277733, + "learning_rate": 0.0006, + "loss": 7.229467391967773, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015362377278506756, + "learning_rate": 0.0006, + "loss": 7.180346488952637, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010505498386919498, + "learning_rate": 0.0006, + "loss": 7.188833236694336, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.009129747748374939, + "learning_rate": 0.0006, + "loss": 7.203153133392334, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.01147883478552103, + "learning_rate": 0.0006, + "loss": 7.163824558258057, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011338942684233189, + "learning_rate": 0.0006, + "loss": 7.16287899017334, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.013277675025165081, + "learning_rate": 0.0006, + "loss": 7.155905723571777, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.015269347466528416, + "learning_rate": 0.0006, + "loss": 7.189371109008789, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.017905596643686295, + "learning_rate": 0.0006, + "loss": 7.177547931671143, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.024431610479950905, + "learning_rate": 0.0006, + "loss": 7.192169189453125, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.03656961768865585, + "learning_rate": 0.0006, + "loss": 7.124387741088867, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.04337048903107643, + "learning_rate": 0.0006, + "loss": 7.140231609344482, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.04047228768467903, + "learning_rate": 0.0006, + "loss": 7.171927452087402, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.07825770974159241, + "learning_rate": 0.0006, + "loss": 7.207631587982178, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.07118063420057297, + "learning_rate": 0.0006, + "loss": 7.209488391876221, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.05153293535113335, + "learning_rate": 0.0006, + "loss": 7.14793062210083, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.03318839520215988, + "learning_rate": 0.0006, + "loss": 7.163010597229004, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.029333539307117462, + "learning_rate": 0.0006, + "loss": 7.159955978393555, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.029405880719423294, + "learning_rate": 0.0006, + "loss": 7.136178493499756, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.037815988063812256, + "learning_rate": 0.0006, + "loss": 7.183186054229736, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.020307250320911407, + "learning_rate": 0.0006, + "loss": 7.133164882659912, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.026048069819808006, + "learning_rate": 0.0006, + "loss": 7.124977111816406, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.027316724881529808, + "learning_rate": 0.0006, + "loss": 7.07265567779541, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.021629009395837784, + "learning_rate": 0.0006, + "loss": 7.150020599365234, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.01605929434299469, + "learning_rate": 0.0006, + "loss": 7.10919189453125, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.020800089463591576, + "learning_rate": 0.0006, + "loss": 7.132045269012451, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.019225774332880974, + "learning_rate": 0.0006, + "loss": 7.128364562988281, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.01561372634023428, + "learning_rate": 0.0006, + "loss": 7.083669662475586, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.021969519555568695, + "learning_rate": 0.0006, + "loss": 7.050841331481934, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.021453695371747017, + "learning_rate": 0.0006, + "loss": 7.097085475921631, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.022836022078990936, + "learning_rate": 0.0006, + "loss": 7.10405969619751, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.02430851384997368, + "learning_rate": 0.0006, + "loss": 7.045181751251221, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.020166993141174316, + "learning_rate": 0.0006, + "loss": 7.068183898925781, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.0118505684658885, + "learning_rate": 0.0006, + "loss": 7.07132625579834, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.023772427812218666, + "learning_rate": 0.0006, + "loss": 7.055238723754883, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.02350712940096855, + "learning_rate": 0.0006, + "loss": 7.055898666381836, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.017244907096028328, + "learning_rate": 0.0006, + "loss": 7.0099711418151855, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.021565575152635574, + "learning_rate": 0.0006, + "loss": 6.991751670837402, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.03241860866546631, + "learning_rate": 0.0006, + "loss": 7.005980491638184, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.04613726586103439, + "learning_rate": 0.0006, + "loss": 7.027859687805176, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.06404894590377808, + "learning_rate": 0.0006, + "loss": 7.123725891113281, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.070647716522217, + "eval_runtime": 58.5801, + "eval_samples_per_second": 41.687, + "eval_steps_per_second": 1.314, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.05542432889342308, + "learning_rate": 0.0006, + "loss": 7.078299522399902, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.026484526693820953, + "learning_rate": 0.0006, + "loss": 6.961159706115723, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.028276391327381134, + "learning_rate": 0.0006, + "loss": 7.019961833953857, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.024486595764756203, + "learning_rate": 0.0006, + "loss": 7.039677619934082, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.017620893195271492, + "learning_rate": 0.0006, + "loss": 6.9270782470703125, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.02547302283346653, + "learning_rate": 0.0006, + "loss": 6.924233436584473, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.02539009042084217, + "learning_rate": 0.0006, + "loss": 6.973979949951172, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.03256227448582649, + "learning_rate": 0.0006, + "loss": 6.933725357055664, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.04651800915598869, + "learning_rate": 0.0006, + "loss": 6.951044082641602, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.04822975769639015, + "learning_rate": 0.0006, + "loss": 6.9614410400390625, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.031165020540356636, + "learning_rate": 0.0006, + "loss": 6.91193151473999, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.029754292219877243, + "learning_rate": 0.0006, + "loss": 6.91802978515625, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.02193152718245983, + "learning_rate": 0.0006, + "loss": 6.928090572357178, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.02428470179438591, + "learning_rate": 0.0006, + "loss": 6.90396785736084, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.023374345153570175, + "learning_rate": 0.0006, + "loss": 6.885858535766602, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.023405000567436218, + "learning_rate": 0.0006, + "loss": 6.8855767250061035, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.017277134582400322, + "learning_rate": 0.0006, + "loss": 6.901449203491211, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.019496750086545944, + "learning_rate": 0.0006, + "loss": 6.885687828063965, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.016893276944756508, + "learning_rate": 0.0006, + "loss": 6.872369766235352, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02058715932071209, + "learning_rate": 0.0006, + "loss": 6.823280334472656, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.015530755743384361, + "learning_rate": 0.0006, + "loss": 6.860499858856201, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.01934926211833954, + "learning_rate": 0.0006, + "loss": 6.817448616027832, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.017413552850484848, + "learning_rate": 0.0006, + "loss": 6.881955146789551, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.026478875428438187, + "learning_rate": 0.0006, + "loss": 6.827404499053955, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.032611701637506485, + "learning_rate": 0.0006, + "loss": 6.78285551071167, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.041900306940078735, + "learning_rate": 0.0006, + "loss": 6.783053874969482, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.04149497672915459, + "learning_rate": 0.0006, + "loss": 6.745220184326172, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.023949826136231422, + "learning_rate": 0.0006, + "loss": 6.841533184051514, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.028601842001080513, + "learning_rate": 0.0006, + "loss": 6.852852821350098, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.027365155518054962, + "learning_rate": 0.0006, + "loss": 6.74376106262207, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.02856568619608879, + "learning_rate": 0.0006, + "loss": 6.74906063079834, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.019172416999936104, + "learning_rate": 0.0006, + "loss": 6.809206008911133, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.02161746844649315, + "learning_rate": 0.0006, + "loss": 6.789888381958008, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.020273666828870773, + "learning_rate": 0.0006, + "loss": 6.73512077331543, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.022329283878207207, + "learning_rate": 0.0006, + "loss": 6.826148509979248, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.02284284122288227, + "learning_rate": 0.0006, + "loss": 6.748521327972412, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.026368247345089912, + "learning_rate": 0.0006, + "loss": 6.645879745483398, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.03627244383096695, + "learning_rate": 0.0006, + "loss": 6.666064739227295, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04140935093164444, + "learning_rate": 0.0006, + "loss": 6.712275981903076, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.0390462800860405, + "learning_rate": 0.0006, + "loss": 6.665900230407715, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.05178583785891533, + "learning_rate": 0.0006, + "loss": 6.753298759460449, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.054328061640262604, + "learning_rate": 0.0006, + "loss": 6.732633113861084, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.04590460658073425, + "learning_rate": 0.0006, + "loss": 6.65250301361084, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.044331666082143784, + "learning_rate": 0.0006, + "loss": 6.68654203414917, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.05308730527758598, + "learning_rate": 0.0006, + "loss": 6.642482757568359, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.03999016433954239, + "learning_rate": 0.0006, + "loss": 6.625227451324463, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.04088086634874344, + "learning_rate": 0.0006, + "loss": 6.67623233795166, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.03615636005997658, + "learning_rate": 0.0006, + "loss": 6.6869001388549805, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.02817360684275627, + "learning_rate": 0.0006, + "loss": 6.674601078033447, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.032586123794317245, + "learning_rate": 0.0006, + "loss": 6.612994194030762, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.031117867678403854, + "learning_rate": 0.0006, + "loss": 6.6244354248046875, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.028330687433481216, + "learning_rate": 0.0006, + "loss": 6.669557094573975, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.021489202976226807, + "learning_rate": 0.0006, + "loss": 6.632035732269287, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.024269424378871918, + "learning_rate": 0.0006, + "loss": 6.660867691040039, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.02237873338162899, + "learning_rate": 0.0006, + "loss": 6.602587699890137, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.026503929868340492, + "learning_rate": 0.0006, + "loss": 6.599200248718262, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.03687124326825142, + "learning_rate": 0.0006, + "loss": 6.567296504974365, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.04708952456712723, + "learning_rate": 0.0006, + "loss": 6.621265411376953, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.041846614331007004, + "learning_rate": 0.0006, + "loss": 6.550536155700684, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.031249675899744034, + "learning_rate": 0.0006, + "loss": 6.54278564453125, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.023082256317138672, + "learning_rate": 0.0006, + "loss": 6.544787406921387, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.020995570346713066, + "learning_rate": 0.0006, + "loss": 6.518143653869629, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020839324221014977, + "learning_rate": 0.0006, + "loss": 6.546667098999023, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.018377432599663734, + "learning_rate": 0.0006, + "loss": 6.478307723999023, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.021713724359869957, + "learning_rate": 0.0006, + "loss": 6.485574245452881, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01359301246702671, + "learning_rate": 0.0006, + "loss": 6.570268630981445, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.016233332455158234, + "learning_rate": 0.0006, + "loss": 6.57150936126709, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.020687608048319817, + "learning_rate": 0.0006, + "loss": 6.527956008911133, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.02032964862883091, + "learning_rate": 0.0006, + "loss": 6.468169689178467, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.016676288098096848, + "learning_rate": 0.0006, + "loss": 6.448418617248535, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.0168539360165596, + "learning_rate": 0.0006, + "loss": 6.461180686950684, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.02666233666241169, + "learning_rate": 0.0006, + "loss": 6.344893932342529, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.5384979248046875, + "eval_runtime": 58.814, + "eval_samples_per_second": 41.521, + "eval_steps_per_second": 1.309, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.04154708981513977, + "learning_rate": 0.0006, + "loss": 6.41855525970459, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.04895668104290962, + "learning_rate": 0.0006, + "loss": 6.499302864074707, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.034464482218027115, + "learning_rate": 0.0006, + "loss": 6.438611030578613, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.06845773756504059, + "learning_rate": 0.0006, + "loss": 6.538361072540283, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.053768858313560486, + "learning_rate": 0.0006, + "loss": 6.527524948120117, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.039037931710481644, + "learning_rate": 0.0006, + "loss": 6.503847122192383, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.03962196782231331, + "learning_rate": 0.0006, + "loss": 6.475178241729736, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.049101535230875015, + "learning_rate": 0.0006, + "loss": 6.474587917327881, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.02830282226204872, + "learning_rate": 0.0006, + "loss": 6.45065975189209, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.023153482005000114, + "learning_rate": 0.0006, + "loss": 6.5218706130981445, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.025064002722501755, + "learning_rate": 0.0006, + "loss": 6.456475257873535, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.02357092685997486, + "learning_rate": 0.0006, + "loss": 6.485952377319336, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.021832305938005447, + "learning_rate": 0.0006, + "loss": 6.425506591796875, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.0227900929749012, + "learning_rate": 0.0006, + "loss": 6.439155578613281, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.015961607918143272, + "learning_rate": 0.0006, + "loss": 6.460862636566162, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.017876390367746353, + "learning_rate": 0.0006, + "loss": 6.44564151763916, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.013703204691410065, + "learning_rate": 0.0006, + "loss": 6.418414115905762, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.015240306034684181, + "learning_rate": 0.0006, + "loss": 6.411238670349121, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01587662845849991, + "learning_rate": 0.0006, + "loss": 6.395642280578613, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.014650700613856316, + "learning_rate": 0.0006, + "loss": 6.302793025970459, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.016533225774765015, + "learning_rate": 0.0006, + "loss": 6.380588054656982, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.019963741302490234, + "learning_rate": 0.0006, + "loss": 6.345336437225342, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.02237936295568943, + "learning_rate": 0.0006, + "loss": 6.387770652770996, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.026658328250050545, + "learning_rate": 0.0006, + "loss": 6.323662757873535, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.02852284163236618, + "learning_rate": 0.0006, + "loss": 6.382204055786133, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.028815090656280518, + "learning_rate": 0.0006, + "loss": 6.402390480041504, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.029393529519438744, + "learning_rate": 0.0006, + "loss": 6.280278205871582, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.02186041697859764, + "learning_rate": 0.0006, + "loss": 6.302707672119141, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.02071218006312847, + "learning_rate": 0.0006, + "loss": 6.3365020751953125, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02779117226600647, + "learning_rate": 0.0006, + "loss": 6.252157211303711, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.03470654785633087, + "learning_rate": 0.0006, + "loss": 6.319070339202881, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.037067804485559464, + "learning_rate": 0.0006, + "loss": 6.314116477966309, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.032182756811380386, + "learning_rate": 0.0006, + "loss": 6.267298698425293, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.026305217295885086, + "learning_rate": 0.0006, + "loss": 6.331688404083252, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.027228357270359993, + "learning_rate": 0.0006, + "loss": 6.31139612197876, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.025781169533729553, + "learning_rate": 0.0006, + "loss": 6.194684982299805, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.029186977073550224, + "learning_rate": 0.0006, + "loss": 6.31306266784668, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.023608777672052383, + "learning_rate": 0.0006, + "loss": 6.285243034362793, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.02267594076693058, + "learning_rate": 0.0006, + "loss": 6.30342960357666, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01926310732960701, + "learning_rate": 0.0006, + "loss": 6.280606269836426, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.026484837755560875, + "learning_rate": 0.0006, + "loss": 6.237980365753174, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.026228854432702065, + "learning_rate": 0.0006, + "loss": 6.3018293380737305, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.022096967324614525, + "learning_rate": 0.0006, + "loss": 6.253863334655762, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.027223890647292137, + "learning_rate": 0.0006, + "loss": 6.166882038116455, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.03160124272108078, + "learning_rate": 0.0006, + "loss": 6.230309963226318, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.04110539332032204, + "learning_rate": 0.0006, + "loss": 6.2334818840026855, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.04355933889746666, + "learning_rate": 0.0006, + "loss": 6.160956382751465, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.03611086308956146, + "learning_rate": 0.0006, + "loss": 6.23399543762207, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.051811400800943375, + "learning_rate": 0.0006, + "loss": 6.241855144500732, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.057434193789958954, + "learning_rate": 0.0006, + "loss": 6.220993995666504, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.04420953616499901, + "learning_rate": 0.0006, + "loss": 6.300461292266846, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.03515457361936569, + "learning_rate": 0.0006, + "loss": 6.269349098205566, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.03222600743174553, + "learning_rate": 0.0006, + "loss": 6.248613357543945, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0384046845138073, + "learning_rate": 0.0006, + "loss": 6.266979217529297, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.0514867827296257, + "learning_rate": 0.0006, + "loss": 6.234103202819824, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.050747793167829514, + "learning_rate": 0.0006, + "loss": 6.322582244873047, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.040878988802433014, + "learning_rate": 0.0006, + "loss": 6.212509632110596, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.034655820578336716, + "learning_rate": 0.0006, + "loss": 6.265879154205322, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.04202224686741829, + "learning_rate": 0.0006, + "loss": 6.2896409034729, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.03632277995347977, + "learning_rate": 0.0006, + "loss": 6.253917217254639, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.028645765036344528, + "learning_rate": 0.0006, + "loss": 6.207965850830078, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.025305170565843582, + "learning_rate": 0.0006, + "loss": 6.168068885803223, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.022137803956866264, + "learning_rate": 0.0006, + "loss": 6.235283374786377, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.01814538985490799, + "learning_rate": 0.0006, + "loss": 6.259641647338867, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.018465086817741394, + "learning_rate": 0.0006, + "loss": 6.111738681793213, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.018638933077454567, + "learning_rate": 0.0006, + "loss": 6.225642204284668, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.016461260616779327, + "learning_rate": 0.0006, + "loss": 6.140647888183594, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.015458385460078716, + "learning_rate": 0.0006, + "loss": 6.155224800109863, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.014934048056602478, + "learning_rate": 0.0006, + "loss": 6.101555824279785, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.013979545794427395, + "learning_rate": 0.0006, + "loss": 6.161718845367432, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.01640394888818264, + "learning_rate": 0.0006, + "loss": 6.055768966674805, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.015135680325329304, + "learning_rate": 0.0006, + "loss": 6.169695854187012, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.16408109664917, + "eval_runtime": 59.3649, + "eval_samples_per_second": 41.135, + "eval_steps_per_second": 1.297, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.012849048711359501, + "learning_rate": 0.0006, + "loss": 6.055830001831055, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.018169576302170753, + "learning_rate": 0.0006, + "loss": 6.082655429840088, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.024293430149555206, + "learning_rate": 0.0006, + "loss": 6.109032154083252, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.03266787901520729, + "learning_rate": 0.0006, + "loss": 6.030613899230957, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03572266176342964, + "learning_rate": 0.0006, + "loss": 6.050390720367432, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.032625213265419006, + "learning_rate": 0.0006, + "loss": 6.169313430786133, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.04051872715353966, + "learning_rate": 0.0006, + "loss": 6.10398006439209, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.0351213738322258, + "learning_rate": 0.0006, + "loss": 6.121973514556885, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03523759916424751, + "learning_rate": 0.0006, + "loss": 6.0626654624938965, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.02706182189285755, + "learning_rate": 0.0006, + "loss": 6.049354076385498, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.024212589487433434, + "learning_rate": 0.0006, + "loss": 6.034826278686523, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024798082187771797, + "learning_rate": 0.0006, + "loss": 6.004058837890625, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.02325567416846752, + "learning_rate": 0.0006, + "loss": 5.986461639404297, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.019060570746660233, + "learning_rate": 0.0006, + "loss": 6.059736251831055, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.016822976991534233, + "learning_rate": 0.0006, + "loss": 6.072957515716553, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.017218658700585365, + "learning_rate": 0.0006, + "loss": 6.031739234924316, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014889383688569069, + "learning_rate": 0.0006, + "loss": 6.025674343109131, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01708107627928257, + "learning_rate": 0.0006, + "loss": 6.090451240539551, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.020260317251086235, + "learning_rate": 0.0006, + "loss": 6.094257354736328, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020110400393605232, + "learning_rate": 0.0006, + "loss": 6.021188735961914, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.02027830481529236, + "learning_rate": 0.0006, + "loss": 6.022156238555908, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.022747062146663666, + "learning_rate": 0.0006, + "loss": 6.011836051940918, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.01990230567753315, + "learning_rate": 0.0006, + "loss": 6.003054618835449, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01596238650381565, + "learning_rate": 0.0006, + "loss": 5.991410732269287, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.015847105532884598, + "learning_rate": 0.0006, + "loss": 6.100622177124023, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.016179397702217102, + "learning_rate": 0.0006, + "loss": 6.032659530639648, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.018256602808833122, + "learning_rate": 0.0006, + "loss": 5.961983680725098, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02805912122130394, + "learning_rate": 0.0006, + "loss": 5.983541965484619, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.039082255214452744, + "learning_rate": 0.0006, + "loss": 6.021474361419678, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.036757659167051315, + "learning_rate": 0.0006, + "loss": 6.005046367645264, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.035277366638183594, + "learning_rate": 0.0006, + "loss": 6.036296844482422, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.034404635429382324, + "learning_rate": 0.0006, + "loss": 5.994539260864258, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.0377206988632679, + "learning_rate": 0.0006, + "loss": 5.941534519195557, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.0389922633767128, + "learning_rate": 0.0006, + "loss": 5.983644485473633, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.04176778718829155, + "learning_rate": 0.0006, + "loss": 6.030922889709473, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.0337153784930706, + "learning_rate": 0.0006, + "loss": 5.961367130279541, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.036116816103458405, + "learning_rate": 0.0006, + "loss": 5.819280624389648, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.030724041163921356, + "learning_rate": 0.0006, + "loss": 5.906380653381348, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.030264202505350113, + "learning_rate": 0.0006, + "loss": 5.975833415985107, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.032096318900585175, + "learning_rate": 0.0006, + "loss": 6.001348972320557, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030579356476664543, + "learning_rate": 0.0006, + "loss": 5.928768634796143, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.028241973370313644, + "learning_rate": 0.0006, + "loss": 5.92582893371582, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.023086953908205032, + "learning_rate": 0.0006, + "loss": 5.892926216125488, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.02547992393374443, + "learning_rate": 0.0006, + "loss": 5.903195381164551, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.023089831694960594, + "learning_rate": 0.0006, + "loss": 5.912033557891846, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.025446368381381035, + "learning_rate": 0.0006, + "loss": 5.926138401031494, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.03748747706413269, + "learning_rate": 0.0006, + "loss": 5.913451194763184, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.03493810072541237, + "learning_rate": 0.0006, + "loss": 5.97633695602417, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.023131275549530983, + "learning_rate": 0.0006, + "loss": 5.899571418762207, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.02986014075577259, + "learning_rate": 0.0006, + "loss": 5.894903182983398, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.030171144753694534, + "learning_rate": 0.0006, + "loss": 5.876595497131348, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.029377546161413193, + "learning_rate": 0.0006, + "loss": 5.911637306213379, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.029829490929841995, + "learning_rate": 0.0006, + "loss": 5.819629669189453, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.024139299988746643, + "learning_rate": 0.0006, + "loss": 5.81728458404541, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.021453432738780975, + "learning_rate": 0.0006, + "loss": 5.987326145172119, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.017942246049642563, + "learning_rate": 0.0006, + "loss": 5.893008232116699, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.019724637269973755, + "learning_rate": 0.0006, + "loss": 5.912441253662109, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.018597912043333054, + "learning_rate": 0.0006, + "loss": 5.860394477844238, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.016232412308454514, + "learning_rate": 0.0006, + "loss": 5.94991397857666, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.01667204685509205, + "learning_rate": 0.0006, + "loss": 5.8506364822387695, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.014232151210308075, + "learning_rate": 0.0006, + "loss": 5.928424835205078, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.01570476032793522, + "learning_rate": 0.0006, + "loss": 5.953692436218262, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.01586179807782173, + "learning_rate": 0.0006, + "loss": 5.820014476776123, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.014729145914316177, + "learning_rate": 0.0006, + "loss": 5.860154628753662, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.013111459091305733, + "learning_rate": 0.0006, + "loss": 5.85988187789917, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.012937922962009907, + "learning_rate": 0.0006, + "loss": 5.756265640258789, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.01377453189343214, + "learning_rate": 0.0006, + "loss": 5.8333048820495605, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.014054795727133751, + "learning_rate": 0.0006, + "loss": 5.803333759307861, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01674959622323513, + "learning_rate": 0.0006, + "loss": 5.8718485832214355, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.018733017146587372, + "learning_rate": 0.0006, + "loss": 5.875979900360107, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.02088983915746212, + "learning_rate": 0.0006, + "loss": 5.763634204864502, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026705985888838768, + "learning_rate": 0.0006, + "loss": 5.6922478675842285, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.847958087921143, + "eval_runtime": 58.5451, + "eval_samples_per_second": 41.711, + "eval_steps_per_second": 1.315, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03181544691324234, + "learning_rate": 0.0006, + "loss": 5.786368370056152, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03238112851977348, + "learning_rate": 0.0006, + "loss": 5.810310363769531, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.03916756808757782, + "learning_rate": 0.0006, + "loss": 5.7761993408203125, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.044009730219841, + "learning_rate": 0.0006, + "loss": 5.898112773895264, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.045235246419906616, + "learning_rate": 0.0006, + "loss": 5.876371383666992, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.050218384712934494, + "learning_rate": 0.0006, + "loss": 5.805103302001953, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.05444490164518356, + "learning_rate": 0.0006, + "loss": 5.826424598693848, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04890982061624527, + "learning_rate": 0.0006, + "loss": 5.7735395431518555, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05478639155626297, + "learning_rate": 0.0006, + "loss": 5.823677062988281, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.057562313973903656, + "learning_rate": 0.0006, + "loss": 5.889334678649902, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.06447703391313553, + "learning_rate": 0.0006, + "loss": 5.84621524810791, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.048861872404813766, + "learning_rate": 0.0006, + "loss": 5.931595802307129, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.05521339178085327, + "learning_rate": 0.0006, + "loss": 5.8229780197143555, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.053666990250349045, + "learning_rate": 0.0006, + "loss": 5.879191875457764, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.0451025515794754, + "learning_rate": 0.0006, + "loss": 5.801628112792969, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.04291655868291855, + "learning_rate": 0.0006, + "loss": 5.914680480957031, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.035102490335702896, + "learning_rate": 0.0006, + "loss": 5.843682765960693, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.03403995931148529, + "learning_rate": 0.0006, + "loss": 5.827154159545898, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.03444375470280647, + "learning_rate": 0.0006, + "loss": 5.800136566162109, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.03165159001946449, + "learning_rate": 0.0006, + "loss": 5.906252384185791, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.026153093203902245, + "learning_rate": 0.0006, + "loss": 5.7423200607299805, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.024357657879590988, + "learning_rate": 0.0006, + "loss": 5.701364517211914, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.021508049219846725, + "learning_rate": 0.0006, + "loss": 5.8328094482421875, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.017313921824097633, + "learning_rate": 0.0006, + "loss": 5.792145252227783, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.018563397228717804, + "learning_rate": 0.0006, + "loss": 5.732672214508057, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.016568679362535477, + "learning_rate": 0.0006, + "loss": 5.81948184967041, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.014133770950138569, + "learning_rate": 0.0006, + "loss": 5.734982490539551, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.014114447869360447, + "learning_rate": 0.0006, + "loss": 5.70073127746582, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.012907393276691437, + "learning_rate": 0.0006, + "loss": 5.758626937866211, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.012673444114625454, + "learning_rate": 0.0006, + "loss": 5.775138854980469, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.01311410591006279, + "learning_rate": 0.0006, + "loss": 5.70920991897583, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.012935544364154339, + "learning_rate": 0.0006, + "loss": 5.781479835510254, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01439738366752863, + "learning_rate": 0.0006, + "loss": 5.691315650939941, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.01205186452716589, + "learning_rate": 0.0006, + "loss": 5.662499904632568, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.011575652286410332, + "learning_rate": 0.0006, + "loss": 5.6271772384643555, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.011672502383589745, + "learning_rate": 0.0006, + "loss": 5.761662006378174, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.011541751213371754, + "learning_rate": 0.0006, + "loss": 5.762078285217285, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.011396365240216255, + "learning_rate": 0.0006, + "loss": 5.67873477935791, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.010688499547541142, + "learning_rate": 0.0006, + "loss": 5.717051982879639, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.012224317528307438, + "learning_rate": 0.0006, + "loss": 5.707948684692383, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.011856825090944767, + "learning_rate": 0.0006, + "loss": 5.70878791809082, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01199064590036869, + "learning_rate": 0.0006, + "loss": 5.708697319030762, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01219446212053299, + "learning_rate": 0.0006, + "loss": 5.6061577796936035, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.013397484086453915, + "learning_rate": 0.0006, + "loss": 5.624789714813232, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01483136136084795, + "learning_rate": 0.0006, + "loss": 5.721141338348389, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.0194488987326622, + "learning_rate": 0.0006, + "loss": 5.604279518127441, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.02079896256327629, + "learning_rate": 0.0006, + "loss": 5.586322784423828, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.018996229395270348, + "learning_rate": 0.0006, + "loss": 5.5987772941589355, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.015779603272676468, + "learning_rate": 0.0006, + "loss": 5.588602066040039, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.015322973020374775, + "learning_rate": 0.0006, + "loss": 5.686467170715332, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.015282213687896729, + "learning_rate": 0.0006, + "loss": 5.664676666259766, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.016589272767305374, + "learning_rate": 0.0006, + "loss": 5.666739463806152, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.01811421848833561, + "learning_rate": 0.0006, + "loss": 5.647593021392822, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.019652029499411583, + "learning_rate": 0.0006, + "loss": 5.628726959228516, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.02118665538728237, + "learning_rate": 0.0006, + "loss": 5.6580705642700195, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.02237832546234131, + "learning_rate": 0.0006, + "loss": 5.635931968688965, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.023897631093859673, + "learning_rate": 0.0006, + "loss": 5.517949104309082, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.02442227490246296, + "learning_rate": 0.0006, + "loss": 5.600021839141846, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.024675287306308746, + "learning_rate": 0.0006, + "loss": 5.6639909744262695, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.030372392386198044, + "learning_rate": 0.0006, + "loss": 5.582888603210449, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.03237690031528473, + "learning_rate": 0.0006, + "loss": 5.573896884918213, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.03511589393019676, + "learning_rate": 0.0006, + "loss": 5.565878868103027, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.03883938118815422, + "learning_rate": 0.0006, + "loss": 5.58540678024292, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.04175502806901932, + "learning_rate": 0.0006, + "loss": 5.6095194816589355, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.04013441503047943, + "learning_rate": 0.0006, + "loss": 5.654401779174805, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.04618770629167557, + "learning_rate": 0.0006, + "loss": 5.6483473777771, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.04507308453321457, + "learning_rate": 0.0006, + "loss": 5.5856218338012695, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.03151383996009827, + "learning_rate": 0.0006, + "loss": 5.594086647033691, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.026883797720074654, + "learning_rate": 0.0006, + "loss": 5.607676029205322, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.02981836162507534, + "learning_rate": 0.0006, + "loss": 5.563666820526123, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.03233477845788002, + "learning_rate": 0.0006, + "loss": 5.667543411254883, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.029927456751465797, + "learning_rate": 0.0006, + "loss": 5.733482837677002, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.679076671600342, + "eval_runtime": 59.5965, + "eval_samples_per_second": 40.976, + "eval_steps_per_second": 1.292, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.027081597596406937, + "learning_rate": 0.0006, + "loss": 5.634487152099609, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.031959421932697296, + "learning_rate": 0.0006, + "loss": 5.643294811248779, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.02802063524723053, + "learning_rate": 0.0006, + "loss": 5.573239326477051, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.032579705119132996, + "learning_rate": 0.0006, + "loss": 5.627150058746338, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.02503928542137146, + "learning_rate": 0.0006, + "loss": 5.6045308113098145, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02421317622065544, + "learning_rate": 0.0006, + "loss": 5.522153377532959, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.021983426064252853, + "learning_rate": 0.0006, + "loss": 5.632939338684082, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.021933183073997498, + "learning_rate": 0.0006, + "loss": 5.592185974121094, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02287031151354313, + "learning_rate": 0.0006, + "loss": 5.613700866699219, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.02281602844595909, + "learning_rate": 0.0006, + "loss": 5.551383018493652, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.021050360053777695, + "learning_rate": 0.0006, + "loss": 5.55275821685791, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.018299926072359085, + "learning_rate": 0.0006, + "loss": 5.528225421905518, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02024853602051735, + "learning_rate": 0.0006, + "loss": 5.522453784942627, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02135239914059639, + "learning_rate": 0.0006, + "loss": 5.527522563934326, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.023394184187054634, + "learning_rate": 0.0006, + "loss": 5.4535017013549805, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.025952080264687538, + "learning_rate": 0.0006, + "loss": 5.5374674797058105, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.022287718951702118, + "learning_rate": 0.0006, + "loss": 5.493753433227539, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.02018447406589985, + "learning_rate": 0.0006, + "loss": 5.510575294494629, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.02050507813692093, + "learning_rate": 0.0006, + "loss": 5.466026782989502, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.023688802495598793, + "learning_rate": 0.0006, + "loss": 5.485815525054932, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.021078843623399734, + "learning_rate": 0.0006, + "loss": 5.415581703186035, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.018266011029481888, + "learning_rate": 0.0006, + "loss": 5.402815818786621, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.019693493843078613, + "learning_rate": 0.0006, + "loss": 5.504674911499023, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.02031373605132103, + "learning_rate": 0.0006, + "loss": 5.503837585449219, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.017636626958847046, + "learning_rate": 0.0006, + "loss": 5.530580997467041, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.01787244901061058, + "learning_rate": 0.0006, + "loss": 5.450218677520752, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.0170669574290514, + "learning_rate": 0.0006, + "loss": 5.481570243835449, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.01802165061235428, + "learning_rate": 0.0006, + "loss": 5.386394500732422, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.022949472069740295, + "learning_rate": 0.0006, + "loss": 5.424929618835449, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.034213390201330185, + "learning_rate": 0.0006, + "loss": 5.346663475036621, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03847593814134598, + "learning_rate": 0.0006, + "loss": 5.418482780456543, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.029393676668405533, + "learning_rate": 0.0006, + "loss": 5.456090927124023, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.03388667106628418, + "learning_rate": 0.0006, + "loss": 5.572983741760254, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.039690613746643066, + "learning_rate": 0.0006, + "loss": 5.414067268371582, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.03630776330828667, + "learning_rate": 0.0006, + "loss": 5.523739814758301, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.03356870263814926, + "learning_rate": 0.0006, + "loss": 5.444767951965332, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.030812304466962814, + "learning_rate": 0.0006, + "loss": 5.461244106292725, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.03319217637181282, + "learning_rate": 0.0006, + "loss": 5.502161026000977, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.032997481524944305, + "learning_rate": 0.0006, + "loss": 5.4162278175354, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.03364962339401245, + "learning_rate": 0.0006, + "loss": 5.415736198425293, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.0344221405684948, + "learning_rate": 0.0006, + "loss": 5.541967868804932, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.029609503224492073, + "learning_rate": 0.0006, + "loss": 5.440614223480225, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.029055926948785782, + "learning_rate": 0.0006, + "loss": 5.462865352630615, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.02658848837018013, + "learning_rate": 0.0006, + "loss": 5.444467544555664, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.026276404038071632, + "learning_rate": 0.0006, + "loss": 5.462278842926025, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.0286889486014843, + "learning_rate": 0.0006, + "loss": 5.4849534034729, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.026849817484617233, + "learning_rate": 0.0006, + "loss": 5.431473731994629, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.02312396466732025, + "learning_rate": 0.0006, + "loss": 5.45462703704834, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.026197485625743866, + "learning_rate": 0.0006, + "loss": 5.452552318572998, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02747255191206932, + "learning_rate": 0.0006, + "loss": 5.447169780731201, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.028123432770371437, + "learning_rate": 0.0006, + "loss": 5.394400596618652, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.02599870041012764, + "learning_rate": 0.0006, + "loss": 5.398341178894043, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.022171657532453537, + "learning_rate": 0.0006, + "loss": 5.368820667266846, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.022309480234980583, + "learning_rate": 0.0006, + "loss": 5.416772365570068, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.024074165150523186, + "learning_rate": 0.0006, + "loss": 5.359002113342285, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.02636653557419777, + "learning_rate": 0.0006, + "loss": 5.484368324279785, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.02203752100467682, + "learning_rate": 0.0006, + "loss": 5.375498294830322, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.019924819469451904, + "learning_rate": 0.0006, + "loss": 5.334723472595215, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.018755966797471046, + "learning_rate": 0.0006, + "loss": 5.348781108856201, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.016747845336794853, + "learning_rate": 0.0006, + "loss": 5.433079719543457, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.0170395877212286, + "learning_rate": 0.0006, + "loss": 5.346794128417969, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.016074176877737045, + "learning_rate": 0.0006, + "loss": 5.3277740478515625, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.014119806699454784, + "learning_rate": 0.0006, + "loss": 5.483551979064941, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.015270394273102283, + "learning_rate": 0.0006, + "loss": 5.378519535064697, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.015677539631724358, + "learning_rate": 0.0006, + "loss": 5.290210247039795, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.015930423513054848, + "learning_rate": 0.0006, + "loss": 5.427360534667969, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.016641564667224884, + "learning_rate": 0.0006, + "loss": 5.301599979400635, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.018293552100658417, + "learning_rate": 0.0006, + "loss": 5.277889251708984, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.017618799582123756, + "learning_rate": 0.0006, + "loss": 5.297194480895996, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015034242533147335, + "learning_rate": 0.0006, + "loss": 5.42428731918335, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.01617511734366417, + "learning_rate": 0.0006, + "loss": 5.420044898986816, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.01837257295846939, + "learning_rate": 0.0006, + "loss": 5.240387916564941, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.423932075500488, + "eval_runtime": 58.0841, + "eval_samples_per_second": 42.042, + "eval_steps_per_second": 1.326, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.020953809842467308, + "learning_rate": 0.0006, + "loss": 5.4320783615112305, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.0227745920419693, + "learning_rate": 0.0006, + "loss": 5.310197830200195, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.021372603252530098, + "learning_rate": 0.0006, + "loss": 5.339468002319336, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.022401731461286545, + "learning_rate": 0.0006, + "loss": 5.279829978942871, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.023942379280924797, + "learning_rate": 0.0006, + "loss": 5.372323989868164, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.024380534887313843, + "learning_rate": 0.0006, + "loss": 5.253486633300781, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.03246625140309334, + "learning_rate": 0.0006, + "loss": 5.305774211883545, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.030399736016988754, + "learning_rate": 0.0006, + "loss": 5.370976448059082, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.027148913592100143, + "learning_rate": 0.0006, + "loss": 5.368010520935059, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.02925540879368782, + "learning_rate": 0.0006, + "loss": 5.303013801574707, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.02915453538298607, + "learning_rate": 0.0006, + "loss": 5.329667091369629, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.031732626259326935, + "learning_rate": 0.0006, + "loss": 5.3634934425354, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.036166731268167496, + "learning_rate": 0.0006, + "loss": 5.223980903625488, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.035654108971357346, + "learning_rate": 0.0006, + "loss": 5.361059188842773, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.0365324430167675, + "learning_rate": 0.0006, + "loss": 5.336151599884033, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.03578880429267883, + "learning_rate": 0.0006, + "loss": 5.434076309204102, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.035734012722969055, + "learning_rate": 0.0006, + "loss": 5.330893516540527, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.03196857124567032, + "learning_rate": 0.0006, + "loss": 5.306009769439697, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.032042086124420166, + "learning_rate": 0.0006, + "loss": 5.343267917633057, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.03160746395587921, + "learning_rate": 0.0006, + "loss": 5.2353715896606445, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.026689818128943443, + "learning_rate": 0.0006, + "loss": 5.254042625427246, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.02880188450217247, + "learning_rate": 0.0006, + "loss": 5.326833248138428, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.027516381815075874, + "learning_rate": 0.0006, + "loss": 5.342336177825928, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.027875030413269997, + "learning_rate": 0.0006, + "loss": 5.200719833374023, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0268265288323164, + "learning_rate": 0.0006, + "loss": 5.249449729919434, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.024824608117341995, + "learning_rate": 0.0006, + "loss": 5.318952560424805, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.022990627214312553, + "learning_rate": 0.0006, + "loss": 5.244993209838867, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.022804604843258858, + "learning_rate": 0.0006, + "loss": 5.3236846923828125, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.02419872209429741, + "learning_rate": 0.0006, + "loss": 5.255486488342285, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.021952059119939804, + "learning_rate": 0.0006, + "loss": 5.3511552810668945, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.022375497967004776, + "learning_rate": 0.0006, + "loss": 5.294790267944336, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019242815673351288, + "learning_rate": 0.0006, + "loss": 5.252618789672852, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.0173486340790987, + "learning_rate": 0.0006, + "loss": 5.30147647857666, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.01612604409456253, + "learning_rate": 0.0006, + "loss": 5.23960542678833, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.014293976128101349, + "learning_rate": 0.0006, + "loss": 5.119932651519775, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.014453536830842495, + "learning_rate": 0.0006, + "loss": 5.166794776916504, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.014685769565403461, + "learning_rate": 0.0006, + "loss": 5.264822959899902, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.01427740603685379, + "learning_rate": 0.0006, + "loss": 5.278433799743652, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.014864951372146606, + "learning_rate": 0.0006, + "loss": 5.214512348175049, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.01565164513885975, + "learning_rate": 0.0006, + "loss": 5.183066368103027, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.01779816672205925, + "learning_rate": 0.0006, + "loss": 5.163185119628906, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.017254827544093132, + "learning_rate": 0.0006, + "loss": 5.243555068969727, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.01650584116578102, + "learning_rate": 0.0006, + "loss": 5.234884738922119, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.017021115869283676, + "learning_rate": 0.0006, + "loss": 5.12271785736084, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.01773759163916111, + "learning_rate": 0.0006, + "loss": 5.242153167724609, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.015679042786359787, + "learning_rate": 0.0006, + "loss": 5.170779228210449, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.013760549947619438, + "learning_rate": 0.0006, + "loss": 5.238644599914551, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.014571522362530231, + "learning_rate": 0.0006, + "loss": 5.177056312561035, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.016209015622735023, + "learning_rate": 0.0006, + "loss": 5.163750648498535, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.016813941299915314, + "learning_rate": 0.0006, + "loss": 5.183428764343262, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.020985357463359833, + "learning_rate": 0.0006, + "loss": 5.229465007781982, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.02679632417857647, + "learning_rate": 0.0006, + "loss": 5.257368087768555, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.025756070390343666, + "learning_rate": 0.0006, + "loss": 5.253736972808838, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.02643490768969059, + "learning_rate": 0.0006, + "loss": 5.228633403778076, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.03303210437297821, + "learning_rate": 0.0006, + "loss": 5.163540363311768, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030432431027293205, + "learning_rate": 0.0006, + "loss": 5.243169784545898, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.028631288558244705, + "learning_rate": 0.0006, + "loss": 5.166398048400879, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.026188310235738754, + "learning_rate": 0.0006, + "loss": 5.128122329711914, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.025308528915047646, + "learning_rate": 0.0006, + "loss": 5.211284637451172, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.020657729357481003, + "learning_rate": 0.0006, + "loss": 5.161575794219971, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.021901234984397888, + "learning_rate": 0.0006, + "loss": 5.20050048828125, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021219318732619286, + "learning_rate": 0.0006, + "loss": 5.211699485778809, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.0200974028557539, + "learning_rate": 0.0006, + "loss": 5.092503547668457, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.023804882541298866, + "learning_rate": 0.0006, + "loss": 5.216068267822266, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.026088010519742966, + "learning_rate": 0.0006, + "loss": 5.155592918395996, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.02601276896893978, + "learning_rate": 0.0006, + "loss": 5.197238922119141, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.020387211814522743, + "learning_rate": 0.0006, + "loss": 5.1955437660217285, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.019214622676372528, + "learning_rate": 0.0006, + "loss": 5.206346035003662, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.019674314185976982, + "learning_rate": 0.0006, + "loss": 5.219857215881348, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.020318234339356422, + "learning_rate": 0.0006, + "loss": 5.191132545471191, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021428676322102547, + "learning_rate": 0.0006, + "loss": 5.214695930480957, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.023235054686665535, + "learning_rate": 0.0006, + "loss": 5.229091644287109, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.289593696594238, + "eval_runtime": 56.9256, + "eval_samples_per_second": 42.898, + "eval_steps_per_second": 1.353, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.02320289984345436, + "learning_rate": 0.0006, + "loss": 4.983669281005859, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.018521282821893692, + "learning_rate": 0.0006, + "loss": 5.200575828552246, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.0180169939994812, + "learning_rate": 0.0006, + "loss": 5.148033618927002, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.019969860091805458, + "learning_rate": 0.0006, + "loss": 5.099125385284424, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01643305830657482, + "learning_rate": 0.0006, + "loss": 5.092347145080566, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.016022363677620888, + "learning_rate": 0.0006, + "loss": 5.129279613494873, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01657041162252426, + "learning_rate": 0.0006, + "loss": 5.159562587738037, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.017924228683114052, + "learning_rate": 0.0006, + "loss": 5.165497779846191, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.021592361852526665, + "learning_rate": 0.0006, + "loss": 5.204775810241699, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.026924120262265205, + "learning_rate": 0.0006, + "loss": 5.128296852111816, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.023156899958848953, + "learning_rate": 0.0006, + "loss": 5.1255340576171875, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.020670367404818535, + "learning_rate": 0.0006, + "loss": 5.128479957580566, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.025447173044085503, + "learning_rate": 0.0006, + "loss": 5.154821872711182, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.027940068393945694, + "learning_rate": 0.0006, + "loss": 5.143466472625732, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.026847844943404198, + "learning_rate": 0.0006, + "loss": 5.045665740966797, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.02340601570904255, + "learning_rate": 0.0006, + "loss": 5.21933126449585, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.02340371161699295, + "learning_rate": 0.0006, + "loss": 4.942949295043945, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.02221992425620556, + "learning_rate": 0.0006, + "loss": 5.197099208831787, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.023508677259087563, + "learning_rate": 0.0006, + "loss": 5.022897720336914, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.026120394468307495, + "learning_rate": 0.0006, + "loss": 5.087725639343262, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.027273228392004967, + "learning_rate": 0.0006, + "loss": 5.090963840484619, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.03241586685180664, + "learning_rate": 0.0006, + "loss": 5.047842979431152, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.030245667323470116, + "learning_rate": 0.0006, + "loss": 5.104803085327148, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.027698364108800888, + "learning_rate": 0.0006, + "loss": 5.0644097328186035, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.029692554846405983, + "learning_rate": 0.0006, + "loss": 5.111942291259766, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.03328656405210495, + "learning_rate": 0.0006, + "loss": 5.1574554443359375, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.031938180327415466, + "learning_rate": 0.0006, + "loss": 5.190371990203857, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.024858945980668068, + "learning_rate": 0.0006, + "loss": 5.121420860290527, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.023033203557133675, + "learning_rate": 0.0006, + "loss": 5.0974345207214355, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.022083545103669167, + "learning_rate": 0.0006, + "loss": 5.046319961547852, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.0228437427431345, + "learning_rate": 0.0006, + "loss": 5.119963645935059, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.0242351982742548, + "learning_rate": 0.0006, + "loss": 5.0444560050964355, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.02401185780763626, + "learning_rate": 0.0006, + "loss": 5.063238620758057, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.02560959942638874, + "learning_rate": 0.0006, + "loss": 5.035224914550781, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.024985626339912415, + "learning_rate": 0.0006, + "loss": 5.049284934997559, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.023178115487098694, + "learning_rate": 0.0006, + "loss": 5.2266621589660645, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.0215512253344059, + "learning_rate": 0.0006, + "loss": 5.1875715255737305, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.01662837713956833, + "learning_rate": 0.0006, + "loss": 5.137012481689453, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.01952764391899109, + "learning_rate": 0.0006, + "loss": 5.072500228881836, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017521077767014503, + "learning_rate": 0.0006, + "loss": 5.045711994171143, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.017355265095829964, + "learning_rate": 0.0006, + "loss": 5.02881383895874, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.015587719157338142, + "learning_rate": 0.0006, + "loss": 5.132787704467773, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.017469845712184906, + "learning_rate": 0.0006, + "loss": 5.105447292327881, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.019630100578069687, + "learning_rate": 0.0006, + "loss": 5.032464981079102, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.01784994639456272, + "learning_rate": 0.0006, + "loss": 5.081345558166504, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.01872754842042923, + "learning_rate": 0.0006, + "loss": 4.9772114753723145, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.01999032311141491, + "learning_rate": 0.0006, + "loss": 5.089740753173828, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.01710767112672329, + "learning_rate": 0.0006, + "loss": 4.9869890213012695, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.015144161880016327, + "learning_rate": 0.0006, + "loss": 5.022564888000488, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.01618543453514576, + "learning_rate": 0.0006, + "loss": 5.035921573638916, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.014653387479484081, + "learning_rate": 0.0006, + "loss": 4.914456367492676, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.013434977270662785, + "learning_rate": 0.0006, + "loss": 5.063764572143555, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.015128864906728268, + "learning_rate": 0.0006, + "loss": 5.052367687225342, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.014210768043994904, + "learning_rate": 0.0006, + "loss": 5.016489028930664, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.015191680751740932, + "learning_rate": 0.0006, + "loss": 4.990176200866699, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.018285127356648445, + "learning_rate": 0.0006, + "loss": 5.015164375305176, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.0204079058021307, + "learning_rate": 0.0006, + "loss": 5.091142654418945, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.01980586163699627, + "learning_rate": 0.0006, + "loss": 5.03350830078125, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.02048729732632637, + "learning_rate": 0.0006, + "loss": 5.03611421585083, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.02058163844048977, + "learning_rate": 0.0006, + "loss": 4.9613447189331055, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.018004924058914185, + "learning_rate": 0.0006, + "loss": 4.966541767120361, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.0198152307420969, + "learning_rate": 0.0006, + "loss": 5.01750373840332, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.018532969057559967, + "learning_rate": 0.0006, + "loss": 5.097424507141113, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018939971923828125, + "learning_rate": 0.0006, + "loss": 5.177223205566406, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.017969170585274696, + "learning_rate": 0.0006, + "loss": 5.118013858795166, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.018021270632743835, + "learning_rate": 0.0006, + "loss": 5.034787178039551, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.017009710893034935, + "learning_rate": 0.0006, + "loss": 5.003920555114746, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.017722049728035927, + "learning_rate": 0.0006, + "loss": 5.031939506530762, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.017134087160229683, + "learning_rate": 0.0006, + "loss": 5.093474388122559, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.018511613830924034, + "learning_rate": 0.0006, + "loss": 5.078604698181152, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.021206334233283997, + "learning_rate": 0.0006, + "loss": 5.052572727203369, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.02239062264561653, + "learning_rate": 0.0006, + "loss": 5.063547134399414, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.173044681549072, + "eval_runtime": 56.7287, + "eval_samples_per_second": 43.047, + "eval_steps_per_second": 1.357, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.020757969468832016, + "learning_rate": 0.0006, + "loss": 5.045648574829102, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.022697702050209045, + "learning_rate": 0.0006, + "loss": 4.993833065032959, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.02434331737458706, + "learning_rate": 0.0006, + "loss": 5.073219299316406, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.025499660521745682, + "learning_rate": 0.0006, + "loss": 5.019072532653809, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.029281053692102432, + "learning_rate": 0.0006, + "loss": 4.959851264953613, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.033878110349178314, + "learning_rate": 0.0006, + "loss": 5.132717609405518, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.03385764732956886, + "learning_rate": 0.0006, + "loss": 5.0482892990112305, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.03466648980975151, + "learning_rate": 0.0006, + "loss": 4.946855545043945, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.028546305373311043, + "learning_rate": 0.0006, + "loss": 5.027774333953857, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.030575869604945183, + "learning_rate": 0.0006, + "loss": 4.9768476486206055, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.03614303469657898, + "learning_rate": 0.0006, + "loss": 5.078113555908203, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.040991660207509995, + "learning_rate": 0.0006, + "loss": 5.033185958862305, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.04551699757575989, + "learning_rate": 0.0006, + "loss": 5.0820746421813965, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.040193989872932434, + "learning_rate": 0.0006, + "loss": 5.059700965881348, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.035851605236530304, + "learning_rate": 0.0006, + "loss": 5.142467498779297, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.034200169146060944, + "learning_rate": 0.0006, + "loss": 5.072312355041504, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.033621110022068024, + "learning_rate": 0.0006, + "loss": 5.08037805557251, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.037372369319200516, + "learning_rate": 0.0006, + "loss": 4.998425483703613, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.037044707685709, + "learning_rate": 0.0006, + "loss": 4.923993110656738, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.030454808846116066, + "learning_rate": 0.0006, + "loss": 5.0889997482299805, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.030969638377428055, + "learning_rate": 0.0006, + "loss": 4.984112739562988, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.0276983380317688, + "learning_rate": 0.0006, + "loss": 4.955412864685059, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.02716052532196045, + "learning_rate": 0.0006, + "loss": 5.0098876953125, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.02403552085161209, + "learning_rate": 0.0006, + "loss": 5.165160655975342, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.0257862601429224, + "learning_rate": 0.0006, + "loss": 5.001547813415527, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.02300378680229187, + "learning_rate": 0.0006, + "loss": 5.038888931274414, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019766854122281075, + "learning_rate": 0.0006, + "loss": 4.924291610717773, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.01888395845890045, + "learning_rate": 0.0006, + "loss": 5.059971332550049, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.015577499754726887, + "learning_rate": 0.0006, + "loss": 4.987686634063721, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016012346372008324, + "learning_rate": 0.0006, + "loss": 5.015157699584961, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.016277998685836792, + "learning_rate": 0.0006, + "loss": 5.004931449890137, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.017504561692476273, + "learning_rate": 0.0006, + "loss": 4.902827739715576, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.01569022797048092, + "learning_rate": 0.0006, + "loss": 4.954863548278809, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.014114422723650932, + "learning_rate": 0.0006, + "loss": 5.00656795501709, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.01573832333087921, + "learning_rate": 0.0006, + "loss": 4.978353023529053, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016480036079883575, + "learning_rate": 0.0006, + "loss": 4.953327655792236, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.014008025638759136, + "learning_rate": 0.0006, + "loss": 5.050840377807617, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.013147998601198196, + "learning_rate": 0.0006, + "loss": 4.974964141845703, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.013807603158056736, + "learning_rate": 0.0006, + "loss": 4.927907943725586, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.013555224984884262, + "learning_rate": 0.0006, + "loss": 4.980690956115723, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.013045408762991428, + "learning_rate": 0.0006, + "loss": 4.9436116218566895, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.012858862057328224, + "learning_rate": 0.0006, + "loss": 4.9595770835876465, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01468253880739212, + "learning_rate": 0.0006, + "loss": 4.90725040435791, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.013224679045379162, + "learning_rate": 0.0006, + "loss": 4.877615928649902, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013555348850786686, + "learning_rate": 0.0006, + "loss": 4.9456610679626465, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.014387160539627075, + "learning_rate": 0.0006, + "loss": 4.9233717918396, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.014354088343679905, + "learning_rate": 0.0006, + "loss": 4.983511924743652, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01332700252532959, + "learning_rate": 0.0006, + "loss": 4.934416770935059, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.011543313041329384, + "learning_rate": 0.0006, + "loss": 4.96388053894043, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.012330878525972366, + "learning_rate": 0.0006, + "loss": 4.946396350860596, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01379779726266861, + "learning_rate": 0.0006, + "loss": 4.981019020080566, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.014612431637942791, + "learning_rate": 0.0006, + "loss": 4.894567966461182, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013062899932265282, + "learning_rate": 0.0006, + "loss": 5.033038139343262, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.013898049481213093, + "learning_rate": 0.0006, + "loss": 4.823636054992676, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.01502019353210926, + "learning_rate": 0.0006, + "loss": 5.020156383514404, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.013810204342007637, + "learning_rate": 0.0006, + "loss": 4.935647010803223, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.013616513460874557, + "learning_rate": 0.0006, + "loss": 4.901486396789551, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.013601024635136127, + "learning_rate": 0.0006, + "loss": 4.931873321533203, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.013668350875377655, + "learning_rate": 0.0006, + "loss": 4.937256813049316, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.01581823080778122, + "learning_rate": 0.0006, + "loss": 4.9611406326293945, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.019478484988212585, + "learning_rate": 0.0006, + "loss": 4.9212236404418945, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.026921333745121956, + "learning_rate": 0.0006, + "loss": 4.88501501083374, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.03155914694070816, + "learning_rate": 0.0006, + "loss": 4.989967346191406, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.029689429327845573, + "learning_rate": 0.0006, + "loss": 4.936000823974609, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.026626063510775566, + "learning_rate": 0.0006, + "loss": 4.984785079956055, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.026168212294578552, + "learning_rate": 0.0006, + "loss": 4.908282279968262, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.02828158251941204, + "learning_rate": 0.0006, + "loss": 4.926390647888184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02649078331887722, + "learning_rate": 0.0006, + "loss": 5.00400972366333, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.0256856270134449, + "learning_rate": 0.0006, + "loss": 5.0317487716674805, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.024728883057832718, + "learning_rate": 0.0006, + "loss": 4.9299635887146, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.026241116225719452, + "learning_rate": 0.0006, + "loss": 4.947319984436035, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.023526743054389954, + "learning_rate": 0.0006, + "loss": 4.94429349899292, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.104345798492432, + "eval_runtime": 57.0955, + "eval_samples_per_second": 42.77, + "eval_steps_per_second": 1.349, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.0216788612306118, + "learning_rate": 0.0006, + "loss": 4.881950855255127, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01830463856458664, + "learning_rate": 0.0006, + "loss": 4.971016883850098, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.02204521745443344, + "learning_rate": 0.0006, + "loss": 4.894623756408691, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.0210255216807127, + "learning_rate": 0.0006, + "loss": 4.834009170532227, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.019797448068857193, + "learning_rate": 0.0006, + "loss": 5.086714744567871, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.02036641724407673, + "learning_rate": 0.0006, + "loss": 4.920056343078613, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.01966066285967827, + "learning_rate": 0.0006, + "loss": 4.879859447479248, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.0202149897813797, + "learning_rate": 0.0006, + "loss": 4.933640480041504, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01893039606511593, + "learning_rate": 0.0006, + "loss": 4.901970863342285, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.016040155664086342, + "learning_rate": 0.0006, + "loss": 4.905491828918457, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01720350608229637, + "learning_rate": 0.0006, + "loss": 4.798816680908203, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.017002852633595467, + "learning_rate": 0.0006, + "loss": 4.851859092712402, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.015594680793583393, + "learning_rate": 0.0006, + "loss": 4.90328311920166, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.014904462732374668, + "learning_rate": 0.0006, + "loss": 4.904215335845947, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.01366228237748146, + "learning_rate": 0.0006, + "loss": 4.9813055992126465, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.014719638973474503, + "learning_rate": 0.0006, + "loss": 4.927306175231934, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.015016036108136177, + "learning_rate": 0.0006, + "loss": 4.874215126037598, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.014925424009561539, + "learning_rate": 0.0006, + "loss": 4.838257789611816, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014335823245346546, + "learning_rate": 0.0006, + "loss": 4.940608024597168, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014987512491643429, + "learning_rate": 0.0006, + "loss": 4.831531524658203, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.016087457537651062, + "learning_rate": 0.0006, + "loss": 4.863227367401123, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.01671300269663334, + "learning_rate": 0.0006, + "loss": 4.8535661697387695, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.015229142270982265, + "learning_rate": 0.0006, + "loss": 4.948635101318359, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01532587967813015, + "learning_rate": 0.0006, + "loss": 4.9035797119140625, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.015715476125478745, + "learning_rate": 0.0006, + "loss": 4.894709587097168, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.014156977646052837, + "learning_rate": 0.0006, + "loss": 4.960206985473633, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.017076566815376282, + "learning_rate": 0.0006, + "loss": 5.005850791931152, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.0172084029763937, + "learning_rate": 0.0006, + "loss": 4.903355598449707, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.0180258359760046, + "learning_rate": 0.0006, + "loss": 4.842667579650879, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.022130120545625687, + "learning_rate": 0.0006, + "loss": 4.799970626831055, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02384897693991661, + "learning_rate": 0.0006, + "loss": 4.860746383666992, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.021881572902202606, + "learning_rate": 0.0006, + "loss": 4.918107032775879, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.023085080087184906, + "learning_rate": 0.0006, + "loss": 4.878946304321289, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.02291012555360794, + "learning_rate": 0.0006, + "loss": 4.958887100219727, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.0199968870729208, + "learning_rate": 0.0006, + "loss": 4.80955696105957, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.018702229484915733, + "learning_rate": 0.0006, + "loss": 5.0276408195495605, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.017787311226129532, + "learning_rate": 0.0006, + "loss": 4.96018123626709, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.01704501546919346, + "learning_rate": 0.0006, + "loss": 4.914553165435791, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.017519677057862282, + "learning_rate": 0.0006, + "loss": 4.798098087310791, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.018511991947889328, + "learning_rate": 0.0006, + "loss": 4.817785263061523, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.01876644790172577, + "learning_rate": 0.0006, + "loss": 4.818995475769043, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.01753568835556507, + "learning_rate": 0.0006, + "loss": 4.811938285827637, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.017046531662344933, + "learning_rate": 0.0006, + "loss": 4.914674282073975, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.01802094094455242, + "learning_rate": 0.0006, + "loss": 4.993284225463867, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.019493764266371727, + "learning_rate": 0.0006, + "loss": 4.895628452301025, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.019461210817098618, + "learning_rate": 0.0006, + "loss": 4.869269371032715, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.01981567218899727, + "learning_rate": 0.0006, + "loss": 4.863052845001221, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.019894491881132126, + "learning_rate": 0.0006, + "loss": 4.8775787353515625, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.01875786855816841, + "learning_rate": 0.0006, + "loss": 4.817763328552246, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.020926695317029953, + "learning_rate": 0.0006, + "loss": 4.7470526695251465, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.021477103233337402, + "learning_rate": 0.0006, + "loss": 4.762682914733887, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.020628711208701134, + "learning_rate": 0.0006, + "loss": 4.807985782623291, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.021489612758159637, + "learning_rate": 0.0006, + "loss": 4.843091011047363, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.021194491535425186, + "learning_rate": 0.0006, + "loss": 4.8300251960754395, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.01833350397646427, + "learning_rate": 0.0006, + "loss": 5.000202178955078, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.018096931278705597, + "learning_rate": 0.0006, + "loss": 4.880648136138916, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.0197161752730608, + "learning_rate": 0.0006, + "loss": 4.788316249847412, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0203181654214859, + "learning_rate": 0.0006, + "loss": 4.870944976806641, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01908455230295658, + "learning_rate": 0.0006, + "loss": 4.890144348144531, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.021281801164150238, + "learning_rate": 0.0006, + "loss": 4.89578914642334, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02124273031949997, + "learning_rate": 0.0006, + "loss": 4.883749008178711, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.020969398319721222, + "learning_rate": 0.0006, + "loss": 4.809142589569092, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.01967657171189785, + "learning_rate": 0.0006, + "loss": 4.821864604949951, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.020887333899736404, + "learning_rate": 0.0006, + "loss": 4.917985916137695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.019203083589673042, + "learning_rate": 0.0006, + "loss": 4.824960708618164, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.020162401720881462, + "learning_rate": 0.0006, + "loss": 4.772617816925049, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.02251487225294113, + "learning_rate": 0.0006, + "loss": 4.735418319702148, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.020313527435064316, + "learning_rate": 0.0006, + "loss": 4.939233779907227, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.021635599434375763, + "learning_rate": 0.0006, + "loss": 4.861807346343994, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.02285311557352543, + "learning_rate": 0.0006, + "loss": 4.932732582092285, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02072535641491413, + "learning_rate": 0.0006, + "loss": 4.796767711639404, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.020588304847478867, + "learning_rate": 0.0006, + "loss": 4.997537612915039, + "step": 864 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.683536033598669e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/l2r50-i2-fulle-lm/checkpoint-864/training_args.bin b/runs/l2r50-i2-fulle-lm/checkpoint-864/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..cad232831c222cc7f8b938b8bbe42f61e7d89c3a --- /dev/null +++ b/runs/l2r50-i2-fulle-lm/checkpoint-864/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c40e3bfa2f5ac1450ca1243e3f2a895d44a06ed5995e6aefd97f78d229d8c8 +size 4792 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/config.json new file mode 100644 index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151674 +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..26beed4f25e71b90d2eaf6a618a168a86d6b91f0 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffe1e08b4029f005c849c88b50a316524a347d2e07c34a943a74a14cad3689ae +size 583362376 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..040ccbd15db63769ecb441ec45ced302b25f3ff9 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4af92d4f36cd57db42725af078d57ed2acd8f24e0b3cffe329864d680577c40 +size 1166837626 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..e2c723a27d937a5234fdd83c600d5494d2b4920c --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce1873f29596200b93f59d5a85747bfece46572d1a0840b7d8fe0df166b999af +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..625005d6c88f4c74f804d5cccc7d953809f2e5f3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:692b8a3b6c6eb68decf07e62e78d5cb6d2f573938e34d6ffe16394958ddf9d90 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..0cbbf074a1f543e1364e32d0b7c6822f36d1bb3c --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35e67c56d39aed6d994eecdadecbd71140c0e5d738c1d60c5f25335076198259 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..b5195c5c2b47ccc1182a69bb7e6216b99388af1f --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:340b9fc530c021cafb0ea29114f89f46734f8bbee4420f160a23216252c1f127 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..630fbc73d70608487940c0c34e65caee42415237 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a67914066d1aebda73626cb9d7d4cbc46ee7df332e5ae85a86cc87be8e519c8 +size 1064 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14 +size 11424648 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/tokenizer_config.json @@ -0,0 +1,21 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>", + "<|mask|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..ea35b40b0f3087e742fa7ab8b4bb7ecb3fbe4950 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/trainer_state.json @@ -0,0 +1,8218 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 16.0, + "eval_steps": 500, + "global_step": 1152, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.5768774151802063, + "learning_rate": 0.0, + "loss": 12.050328254699707, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.5784570574760437, + "learning_rate": 5.999999999999999e-06, + "loss": 12.049251556396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.5607187151908875, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.921792030334473, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.46466919779777527, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.686628341674805, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.38012945652008057, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.432823181152344, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.33215776085853577, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.207542419433594, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.2928776741027832, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.000097274780273, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.26018962264060974, + "learning_rate": 4.2e-05, + "loss": 10.825748443603516, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.23717083036899567, + "learning_rate": 4.7999999999999994e-05, + "loss": 10.677122116088867, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.21641355752944946, + "learning_rate": 5.399999999999999e-05, + "loss": 10.569148063659668, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.1982956826686859, + "learning_rate": 5.9999999999999995e-05, + "loss": 10.49653434753418, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.18819282948970795, + "learning_rate": 6.599999999999999e-05, + "loss": 10.432951927185059, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.18322618305683136, + "learning_rate": 7.199999999999999e-05, + "loss": 10.387115478515625, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.18399226665496826, + "learning_rate": 7.8e-05, + "loss": 10.33381462097168, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.1833190619945526, + "learning_rate": 8.4e-05, + "loss": 10.301689147949219, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.18490834534168243, + "learning_rate": 8.999999999999999e-05, + "loss": 10.240968704223633, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.18358023464679718, + "learning_rate": 9.599999999999999e-05, + "loss": 10.194717407226562, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.1801682710647583, + "learning_rate": 0.000102, + "loss": 10.14897632598877, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.17899440228939056, + "learning_rate": 0.00010799999999999998, + "loss": 10.078763961791992, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.17666497826576233, + "learning_rate": 0.00011399999999999999, + "loss": 10.01095962524414, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.17196154594421387, + "learning_rate": 0.00011999999999999999, + "loss": 9.951203346252441, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.169203519821167, + "learning_rate": 0.00012599999999999997, + "loss": 9.871726036071777, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.16500306129455566, + "learning_rate": 0.00013199999999999998, + "loss": 9.803646087646484, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.16128124296665192, + "learning_rate": 0.000138, + "loss": 9.728487014770508, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.15904058516025543, + "learning_rate": 0.00014399999999999998, + "loss": 9.642473220825195, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.15560561418533325, + "learning_rate": 0.00015, + "loss": 9.56414794921875, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.15278612077236176, + "learning_rate": 0.000156, + "loss": 9.484230041503906, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.15263237059116364, + "learning_rate": 0.000162, + "loss": 9.383644104003906, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.14945024251937866, + "learning_rate": 0.000168, + "loss": 9.29840087890625, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.14614646136760712, + "learning_rate": 0.00017399999999999997, + "loss": 9.208931922912598, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.14295098185539246, + "learning_rate": 0.00017999999999999998, + "loss": 9.122323989868164, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.13868345320224762, + "learning_rate": 0.000186, + "loss": 9.03034496307373, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.135064035654068, + "learning_rate": 0.00019199999999999998, + "loss": 8.92818832397461, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.13097241520881653, + "learning_rate": 0.000198, + "loss": 8.834627151489258, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.12478776276111603, + "learning_rate": 0.000204, + "loss": 8.755172729492188, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.1214982345700264, + "learning_rate": 0.00020999999999999998, + "loss": 8.647891998291016, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.11958328634500504, + "learning_rate": 0.00021599999999999996, + "loss": 8.543274879455566, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.11680915951728821, + "learning_rate": 0.00022199999999999998, + "loss": 8.44540023803711, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.11491474509239197, + "learning_rate": 0.00022799999999999999, + "loss": 8.349921226501465, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.11160755902528763, + "learning_rate": 0.000234, + "loss": 8.26960563659668, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.10840637236833572, + "learning_rate": 0.00023999999999999998, + "loss": 8.192136764526367, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.10585298389196396, + "learning_rate": 0.00024599999999999996, + "loss": 8.09546947479248, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.1032542735338211, + "learning_rate": 0.00025199999999999995, + "loss": 7.991090774536133, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.0986943170428276, + "learning_rate": 0.000258, + "loss": 7.92384147644043, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.09320762753486633, + "learning_rate": 0.00026399999999999997, + "loss": 7.8478193283081055, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.09168332070112228, + "learning_rate": 0.00027, + "loss": 7.76185417175293, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.08832395076751709, + "learning_rate": 0.000276, + "loss": 7.679399490356445, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.08816400915384293, + "learning_rate": 0.00028199999999999997, + "loss": 7.5789642333984375, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0861717164516449, + "learning_rate": 0.00028799999999999995, + "loss": 7.519214630126953, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.08103746175765991, + "learning_rate": 0.000294, + "loss": 7.4475321769714355, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.07545104622840881, + "learning_rate": 0.0003, + "loss": 7.401182174682617, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.07042575627565384, + "learning_rate": 0.00030599999999999996, + "loss": 7.340173244476318, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.06737572699785233, + "learning_rate": 0.000312, + "loss": 7.265697002410889, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.06151146441698074, + "learning_rate": 0.000318, + "loss": 7.231061935424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.05638071522116661, + "learning_rate": 0.000324, + "loss": 7.195466995239258, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.05084463208913803, + "learning_rate": 0.00033, + "loss": 7.137603759765625, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.04293261095881462, + "learning_rate": 0.000336, + "loss": 7.115458965301514, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.036064863204956055, + "learning_rate": 0.00034199999999999996, + "loss": 7.089038848876953, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.029026754200458527, + "learning_rate": 0.00034799999999999995, + "loss": 7.0584306716918945, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.02286517433822155, + "learning_rate": 0.00035399999999999993, + "loss": 7.049450874328613, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.018878433853387833, + "learning_rate": 0.00035999999999999997, + "loss": 7.008824348449707, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.010773813351988792, + "learning_rate": 0.00036599999999999995, + "loss": 7.024829864501953, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.008794732391834259, + "learning_rate": 0.000372, + "loss": 7.018399238586426, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.011709796264767647, + "learning_rate": 0.00037799999999999997, + "loss": 7.0101752281188965, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.012123258784413338, + "learning_rate": 0.00038399999999999996, + "loss": 7.024317741394043, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01530320756137371, + "learning_rate": 0.00039, + "loss": 7.027738571166992, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.01890002004802227, + "learning_rate": 0.000396, + "loss": 7.024099349975586, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.020684150978922844, + "learning_rate": 0.000402, + "loss": 7.007943153381348, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.019826196134090424, + "learning_rate": 0.000408, + "loss": 7.006874084472656, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.019218673929572105, + "learning_rate": 0.0004139999999999999, + "loss": 7.002897262573242, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.013856411911547184, + "learning_rate": 0.00041999999999999996, + "loss": 6.993999481201172, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.010390003211796284, + "learning_rate": 0.00042599999999999995, + "loss": 7.000683784484863, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.91140079498291, + "eval_runtime": 57.7925, + "eval_samples_per_second": 42.255, + "eval_steps_per_second": 1.332, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.013802163302898407, + "learning_rate": 0.00043199999999999993, + "loss": 6.998257637023926, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.010109687224030495, + "learning_rate": 0.00043799999999999997, + "loss": 7.002542018890381, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.008019139990210533, + "learning_rate": 0.00044399999999999995, + "loss": 6.985941410064697, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.0069672465324401855, + "learning_rate": 0.00045, + "loss": 6.99245023727417, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.00714076729491353, + "learning_rate": 0.00045599999999999997, + "loss": 6.959296226501465, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.008408435620367527, + "learning_rate": 0.00046199999999999995, + "loss": 6.954504489898682, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.008778786286711693, + "learning_rate": 0.000468, + "loss": 7.0007004737854, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.007753407116979361, + "learning_rate": 0.000474, + "loss": 6.98183536529541, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.01087260153144598, + "learning_rate": 0.00047999999999999996, + "loss": 6.991066932678223, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.006122548598796129, + "learning_rate": 0.000486, + "loss": 6.978824138641357, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.0067383465357124805, + "learning_rate": 0.0004919999999999999, + "loss": 6.961703777313232, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.007922791875898838, + "learning_rate": 0.000498, + "loss": 6.954860687255859, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.00818922370672226, + "learning_rate": 0.0005039999999999999, + "loss": 6.945961952209473, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.006791461259126663, + "learning_rate": 0.0005099999999999999, + "loss": 6.968680381774902, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.005394163075834513, + "learning_rate": 0.000516, + "loss": 6.956404685974121, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.004500220064073801, + "learning_rate": 0.000522, + "loss": 6.96998405456543, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008699574507772923, + "learning_rate": 0.0005279999999999999, + "loss": 6.955585479736328, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.005889365915209055, + "learning_rate": 0.000534, + "loss": 6.9708051681518555, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009646810591220856, + "learning_rate": 0.00054, + "loss": 6.962608814239502, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.004517171997576952, + "learning_rate": 0.0005459999999999999, + "loss": 6.957627296447754, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.006339677143841982, + "learning_rate": 0.000552, + "loss": 6.963141441345215, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.005387383047491312, + "learning_rate": 0.000558, + "loss": 6.934465408325195, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.005252422299236059, + "learning_rate": 0.0005639999999999999, + "loss": 6.956108093261719, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.008699414320290089, + "learning_rate": 0.00057, + "loss": 6.973369598388672, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.003947409335523844, + "learning_rate": 0.0005759999999999999, + "loss": 6.950225830078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.00768681475892663, + "learning_rate": 0.0005819999999999999, + "loss": 6.964739799499512, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.006563829258084297, + "learning_rate": 0.000588, + "loss": 6.972421169281006, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007606555242091417, + "learning_rate": 0.0005939999999999999, + "loss": 6.958213806152344, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.006250880658626556, + "learning_rate": 0.0006, + "loss": 6.962296485900879, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.0073033408261835575, + "learning_rate": 0.0006, + "loss": 6.969705104827881, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.0054510910995304585, + "learning_rate": 0.0006, + "loss": 6.9652862548828125, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.004895673133432865, + "learning_rate": 0.0006, + "loss": 6.956023693084717, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.005306515377014875, + "learning_rate": 0.0006, + "loss": 6.963231086730957, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.004742146469652653, + "learning_rate": 0.0006, + "loss": 6.9582061767578125, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.004220000468194485, + "learning_rate": 0.0006, + "loss": 6.972961902618408, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0042153168469667435, + "learning_rate": 0.0006, + "loss": 6.961165904998779, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.005165040958672762, + "learning_rate": 0.0006, + "loss": 6.95018196105957, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.003930266480892897, + "learning_rate": 0.0006, + "loss": 6.9651689529418945, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.004465071018785238, + "learning_rate": 0.0006, + "loss": 6.960553169250488, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.004888612311333418, + "learning_rate": 0.0006, + "loss": 6.967258930206299, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.007497863378375769, + "learning_rate": 0.0006, + "loss": 6.954545021057129, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.005659648682922125, + "learning_rate": 0.0006, + "loss": 6.962299823760986, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.006641396787017584, + "learning_rate": 0.0006, + "loss": 6.952983856201172, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.009715928696095943, + "learning_rate": 0.0006, + "loss": 6.958622932434082, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.012789523229002953, + "learning_rate": 0.0006, + "loss": 6.932043075561523, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021673237904906273, + "learning_rate": 0.0006, + "loss": 6.931290149688721, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.16024991869926453, + "learning_rate": 0.0006, + "loss": 6.979497909545898, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.2434523105621338, + "learning_rate": 0.0006, + "loss": 7.185218811035156, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.39252421259880066, + "learning_rate": 0.0006, + "loss": 7.137226581573486, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.10565312206745148, + "learning_rate": 0.0006, + "loss": 6.984494209289551, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.1033431813120842, + "learning_rate": 0.0006, + "loss": 7.001287937164307, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.14875490963459015, + "learning_rate": 0.0006, + "loss": 7.051436901092529, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.1374562531709671, + "learning_rate": 0.0006, + "loss": 7.0293121337890625, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.0910211056470871, + "learning_rate": 0.0006, + "loss": 7.007458686828613, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.021633487194776535, + "learning_rate": 0.0006, + "loss": 6.996816635131836, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.08692754060029984, + "learning_rate": 0.0006, + "loss": 6.998541831970215, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.12613292038440704, + "learning_rate": 0.0006, + "loss": 6.9970703125, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.09462810307741165, + "learning_rate": 0.0006, + "loss": 6.997045040130615, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.03534962609410286, + "learning_rate": 0.0006, + "loss": 6.9738450050354, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.033839885145425797, + "learning_rate": 0.0006, + "loss": 6.946001052856445, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.06263269484043121, + "learning_rate": 0.0006, + "loss": 6.968981742858887, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07117997109889984, + "learning_rate": 0.0006, + "loss": 6.981535911560059, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.06336277723312378, + "learning_rate": 0.0006, + "loss": 6.976946830749512, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047467734664678574, + "learning_rate": 0.0006, + "loss": 6.961673736572266, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.018787868320941925, + "learning_rate": 0.0006, + "loss": 6.948919296264648, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.025909408926963806, + "learning_rate": 0.0006, + "loss": 6.976190567016602, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.052842333912849426, + "learning_rate": 0.0006, + "loss": 6.971207618713379, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.06421888619661331, + "learning_rate": 0.0006, + "loss": 6.979109764099121, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.055596910417079926, + "learning_rate": 0.0006, + "loss": 6.972118854522705, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.03267091140151024, + "learning_rate": 0.0006, + "loss": 6.947260856628418, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.008032613433897495, + "learning_rate": 0.0006, + "loss": 6.936816215515137, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.020184025168418884, + "learning_rate": 0.0006, + "loss": 6.929848670959473, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.836523056030273, + "eval_runtime": 58.5622, + "eval_samples_per_second": 41.699, + "eval_steps_per_second": 1.315, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.031980376690626144, + "learning_rate": 0.0006, + "loss": 6.968081474304199, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03802533075213432, + "learning_rate": 0.0006, + "loss": 6.95728874206543, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03796226158738136, + "learning_rate": 0.0006, + "loss": 6.96516752243042, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.02621951326727867, + "learning_rate": 0.0006, + "loss": 6.956058502197266, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.013292718678712845, + "learning_rate": 0.0006, + "loss": 6.9783935546875, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.009446968324482441, + "learning_rate": 0.0006, + "loss": 6.96873664855957, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.021703092381358147, + "learning_rate": 0.0006, + "loss": 6.959234237670898, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.027765290811657906, + "learning_rate": 0.0006, + "loss": 6.968524932861328, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.028761165216565132, + "learning_rate": 0.0006, + "loss": 6.954620838165283, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.019729383289813995, + "learning_rate": 0.0006, + "loss": 6.949009895324707, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.008990089409053326, + "learning_rate": 0.0006, + "loss": 6.956226348876953, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.00549644511193037, + "learning_rate": 0.0006, + "loss": 6.957143783569336, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.014436670579016209, + "learning_rate": 0.0006, + "loss": 6.954257011413574, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.01990494504570961, + "learning_rate": 0.0006, + "loss": 6.942131519317627, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.019155776128172874, + "learning_rate": 0.0006, + "loss": 6.946969985961914, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.014970783144235611, + "learning_rate": 0.0006, + "loss": 6.957938194274902, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.009875659830868244, + "learning_rate": 0.0006, + "loss": 6.953309059143066, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.004792087245732546, + "learning_rate": 0.0006, + "loss": 6.975146770477295, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.009936527349054813, + "learning_rate": 0.0006, + "loss": 6.961674690246582, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.015100127086043358, + "learning_rate": 0.0006, + "loss": 6.968759536743164, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01449542585760355, + "learning_rate": 0.0006, + "loss": 6.966510772705078, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.01073537115007639, + "learning_rate": 0.0006, + "loss": 6.936373710632324, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.006766272243112326, + "learning_rate": 0.0006, + "loss": 6.96397590637207, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.005055439658463001, + "learning_rate": 0.0006, + "loss": 6.961050987243652, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.006252622697502375, + "learning_rate": 0.0006, + "loss": 6.980506420135498, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.008084346540272236, + "learning_rate": 0.0006, + "loss": 6.9346604347229, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.010260380804538727, + "learning_rate": 0.0006, + "loss": 6.964118957519531, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.010697875171899796, + "learning_rate": 0.0006, + "loss": 6.941572189331055, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.007338830269873142, + "learning_rate": 0.0006, + "loss": 6.950471878051758, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.005931832827627659, + "learning_rate": 0.0006, + "loss": 6.962935924530029, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.0061243255622684956, + "learning_rate": 0.0006, + "loss": 6.946418285369873, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.007983328774571419, + "learning_rate": 0.0006, + "loss": 6.959758758544922, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.010470672510564327, + "learning_rate": 0.0006, + "loss": 6.93851375579834, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.011147388257086277, + "learning_rate": 0.0006, + "loss": 6.93258810043335, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01058544497936964, + "learning_rate": 0.0006, + "loss": 6.932259559631348, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.016850745305418968, + "learning_rate": 0.0006, + "loss": 6.924542427062988, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.023101547732949257, + "learning_rate": 0.0006, + "loss": 6.914878845214844, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.015795735642313957, + "learning_rate": 0.0006, + "loss": 6.922432899475098, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.008821426890790462, + "learning_rate": 0.0006, + "loss": 6.924562454223633, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.012502548284828663, + "learning_rate": 0.0006, + "loss": 6.914718151092529, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.011545066721737385, + "learning_rate": 0.0006, + "loss": 6.915705680847168, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.022190893068909645, + "learning_rate": 0.0006, + "loss": 6.88851261138916, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.023395851254463196, + "learning_rate": 0.0006, + "loss": 6.877655029296875, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015601156279444695, + "learning_rate": 0.0006, + "loss": 6.889644145965576, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01362934336066246, + "learning_rate": 0.0006, + "loss": 6.891547203063965, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.012305490672588348, + "learning_rate": 0.0006, + "loss": 6.890357971191406, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.01623358018696308, + "learning_rate": 0.0006, + "loss": 6.849443435668945, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.018063291907310486, + "learning_rate": 0.0006, + "loss": 6.842372894287109, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.010179434902966022, + "learning_rate": 0.0006, + "loss": 6.870055198669434, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.009986722841858864, + "learning_rate": 0.0006, + "loss": 6.859061241149902, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.01117654051631689, + "learning_rate": 0.0006, + "loss": 6.861199378967285, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.008551654405891895, + "learning_rate": 0.0006, + "loss": 6.850255012512207, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.018904326483607292, + "learning_rate": 0.0006, + "loss": 6.848639965057373, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.02080724574625492, + "learning_rate": 0.0006, + "loss": 6.80456018447876, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.012010873295366764, + "learning_rate": 0.0006, + "loss": 6.848775863647461, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.008113056421279907, + "learning_rate": 0.0006, + "loss": 6.849387168884277, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.012912592850625515, + "learning_rate": 0.0006, + "loss": 6.832240104675293, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.01146865077316761, + "learning_rate": 0.0006, + "loss": 6.850383281707764, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.0171892661601305, + "learning_rate": 0.0006, + "loss": 6.839577674865723, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.012083170004189014, + "learning_rate": 0.0006, + "loss": 6.818614482879639, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.0062555656768381596, + "learning_rate": 0.0006, + "loss": 6.831291198730469, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.010102466680109501, + "learning_rate": 0.0006, + "loss": 6.846522808074951, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.009042087942361832, + "learning_rate": 0.0006, + "loss": 6.822072982788086, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.011133556254208088, + "learning_rate": 0.0006, + "loss": 6.811252593994141, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.00990577507764101, + "learning_rate": 0.0006, + "loss": 6.814435958862305, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.006605073343962431, + "learning_rate": 0.0006, + "loss": 6.831300735473633, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.007831175811588764, + "learning_rate": 0.0006, + "loss": 6.841738700866699, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.005536255426704884, + "learning_rate": 0.0006, + "loss": 6.8043670654296875, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.006532667670398951, + "learning_rate": 0.0006, + "loss": 6.8091840744018555, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.006674565840512514, + "learning_rate": 0.0006, + "loss": 6.811327934265137, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.01150344219058752, + "learning_rate": 0.0006, + "loss": 6.810762405395508, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.009015528485178947, + "learning_rate": 0.0006, + "loss": 6.836431980133057, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.718832969665527, + "eval_runtime": 58.6615, + "eval_samples_per_second": 41.629, + "eval_steps_per_second": 1.313, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.009791829623281956, + "learning_rate": 0.0006, + "loss": 6.8102593421936035, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.008970947004854679, + "learning_rate": 0.0006, + "loss": 6.792663097381592, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.013689392246305943, + "learning_rate": 0.0006, + "loss": 6.815651893615723, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.031499531120061874, + "learning_rate": 0.0006, + "loss": 6.809841156005859, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03203094005584717, + "learning_rate": 0.0006, + "loss": 6.779417037963867, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.007947996258735657, + "learning_rate": 0.0006, + "loss": 6.773656845092773, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03152519091963768, + "learning_rate": 0.0006, + "loss": 6.804083824157715, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.014937258325517178, + "learning_rate": 0.0006, + "loss": 6.792213439941406, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.020961705595254898, + "learning_rate": 0.0006, + "loss": 6.8092803955078125, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.009152988903224468, + "learning_rate": 0.0006, + "loss": 6.7737603187561035, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.02770620957016945, + "learning_rate": 0.0006, + "loss": 6.767014503479004, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.01165871974080801, + "learning_rate": 0.0006, + "loss": 6.760884761810303, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.01910376362502575, + "learning_rate": 0.0006, + "loss": 6.775722980499268, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.010413747280836105, + "learning_rate": 0.0006, + "loss": 6.779816627502441, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.01918417401611805, + "learning_rate": 0.0006, + "loss": 6.760381698608398, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.017960334196686745, + "learning_rate": 0.0006, + "loss": 6.761198997497559, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.016546795144677162, + "learning_rate": 0.0006, + "loss": 6.742448806762695, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.01889597810804844, + "learning_rate": 0.0006, + "loss": 6.758237838745117, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.009611992165446281, + "learning_rate": 0.0006, + "loss": 6.75614070892334, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02034934051334858, + "learning_rate": 0.0006, + "loss": 6.736965179443359, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.010740065947175026, + "learning_rate": 0.0006, + "loss": 6.746947288513184, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.018383340910077095, + "learning_rate": 0.0006, + "loss": 6.747332572937012, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.012362911365926266, + "learning_rate": 0.0006, + "loss": 6.767669200897217, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.01127463299781084, + "learning_rate": 0.0006, + "loss": 6.769808292388916, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.017693769186735153, + "learning_rate": 0.0006, + "loss": 6.733658790588379, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.006420376244932413, + "learning_rate": 0.0006, + "loss": 6.74934720993042, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.014020364731550217, + "learning_rate": 0.0006, + "loss": 6.728157043457031, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.008018582127988338, + "learning_rate": 0.0006, + "loss": 6.730726718902588, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.00782222580164671, + "learning_rate": 0.0006, + "loss": 6.7499895095825195, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.008278409950435162, + "learning_rate": 0.0006, + "loss": 6.712803363800049, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.009826759807765484, + "learning_rate": 0.0006, + "loss": 6.742262840270996, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.009488224983215332, + "learning_rate": 0.0006, + "loss": 6.739130020141602, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.007667903788387775, + "learning_rate": 0.0006, + "loss": 6.747515678405762, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.0073090000078082085, + "learning_rate": 0.0006, + "loss": 6.740870475769043, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.010058708488941193, + "learning_rate": 0.0006, + "loss": 6.777563095092773, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.020676063373684883, + "learning_rate": 0.0006, + "loss": 6.708603858947754, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.027210544794797897, + "learning_rate": 0.0006, + "loss": 6.694097995758057, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.038053520023822784, + "learning_rate": 0.0006, + "loss": 6.697017669677734, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04684312269091606, + "learning_rate": 0.0006, + "loss": 6.729883193969727, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03673060983419418, + "learning_rate": 0.0006, + "loss": 6.743753433227539, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.006882632151246071, + "learning_rate": 0.0006, + "loss": 6.71405029296875, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.03588201105594635, + "learning_rate": 0.0006, + "loss": 6.712467193603516, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.029415199533104897, + "learning_rate": 0.0006, + "loss": 6.715634822845459, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.011425039730966091, + "learning_rate": 0.0006, + "loss": 6.697334289550781, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.0297263003885746, + "learning_rate": 0.0006, + "loss": 6.724848747253418, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.007282314356416464, + "learning_rate": 0.0006, + "loss": 6.702306270599365, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.02673051692545414, + "learning_rate": 0.0006, + "loss": 6.72357177734375, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.016174081712961197, + "learning_rate": 0.0006, + "loss": 6.730698585510254, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.0182416420429945, + "learning_rate": 0.0006, + "loss": 6.717386245727539, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.023250866681337357, + "learning_rate": 0.0006, + "loss": 6.697832107543945, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.007539753336459398, + "learning_rate": 0.0006, + "loss": 6.714348793029785, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.030086630955338478, + "learning_rate": 0.0006, + "loss": 6.724481582641602, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.012784913182258606, + "learning_rate": 0.0006, + "loss": 6.703491687774658, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.02042769268155098, + "learning_rate": 0.0006, + "loss": 6.721139907836914, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.023557329550385475, + "learning_rate": 0.0006, + "loss": 6.728122711181641, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.007716621737927198, + "learning_rate": 0.0006, + "loss": 6.70085334777832, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022772278636693954, + "learning_rate": 0.0006, + "loss": 6.712882995605469, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.007928933948278427, + "learning_rate": 0.0006, + "loss": 6.698320388793945, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.022355245426297188, + "learning_rate": 0.0006, + "loss": 6.687088489532471, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.02037850394845009, + "learning_rate": 0.0006, + "loss": 6.676856994628906, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.012290362268686295, + "learning_rate": 0.0006, + "loss": 6.682883262634277, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.022331275045871735, + "learning_rate": 0.0006, + "loss": 6.694752216339111, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.007527181413024664, + "learning_rate": 0.0006, + "loss": 6.686394691467285, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.02159409038722515, + "learning_rate": 0.0006, + "loss": 6.697943687438965, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.015054763294756413, + "learning_rate": 0.0006, + "loss": 6.677796363830566, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01466019731014967, + "learning_rate": 0.0006, + "loss": 6.697715759277344, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.01875465176999569, + "learning_rate": 0.0006, + "loss": 6.700228691101074, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.00778991961851716, + "learning_rate": 0.0006, + "loss": 6.69331693649292, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.020182441920042038, + "learning_rate": 0.0006, + "loss": 6.672651290893555, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.01240516733378172, + "learning_rate": 0.0006, + "loss": 6.688860893249512, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.011218412779271603, + "learning_rate": 0.0006, + "loss": 6.669534683227539, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.0141852255910635, + "learning_rate": 0.0006, + "loss": 6.690206527709961, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 7.537620544433594, + "eval_runtime": 58.4398, + "eval_samples_per_second": 41.787, + "eval_steps_per_second": 1.318, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.008106146939098835, + "learning_rate": 0.0006, + "loss": 6.688591003417969, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.014588244259357452, + "learning_rate": 0.0006, + "loss": 6.684961318969727, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.011985368095338345, + "learning_rate": 0.0006, + "loss": 6.672794342041016, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.014632102102041245, + "learning_rate": 0.0006, + "loss": 6.702736854553223, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.011243713088333607, + "learning_rate": 0.0006, + "loss": 6.658792495727539, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.010519202798604965, + "learning_rate": 0.0006, + "loss": 6.662495136260986, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.007318088319152594, + "learning_rate": 0.0006, + "loss": 6.675954818725586, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.011712398380041122, + "learning_rate": 0.0006, + "loss": 6.673973083496094, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.006726635154336691, + "learning_rate": 0.0006, + "loss": 6.656833171844482, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.005466983187943697, + "learning_rate": 0.0006, + "loss": 6.675468921661377, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.007163621485233307, + "learning_rate": 0.0006, + "loss": 6.6862874031066895, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.007626629434525967, + "learning_rate": 0.0006, + "loss": 6.654741287231445, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.0066330647096037865, + "learning_rate": 0.0006, + "loss": 6.68114709854126, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.007363482844084501, + "learning_rate": 0.0006, + "loss": 6.683980941772461, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.00971233006566763, + "learning_rate": 0.0006, + "loss": 6.669700622558594, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.013185256160795689, + "learning_rate": 0.0006, + "loss": 6.686490535736084, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.019116027280688286, + "learning_rate": 0.0006, + "loss": 6.645479202270508, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0328814722597599, + "learning_rate": 0.0006, + "loss": 6.68080997467041, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04482287913560867, + "learning_rate": 0.0006, + "loss": 6.698016166687012, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.03053313121199608, + "learning_rate": 0.0006, + "loss": 6.642240524291992, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.043351784348487854, + "learning_rate": 0.0006, + "loss": 6.662343978881836, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.03165886923670769, + "learning_rate": 0.0006, + "loss": 6.6291961669921875, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.028601787984371185, + "learning_rate": 0.0006, + "loss": 6.667544364929199, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.019558480009436607, + "learning_rate": 0.0006, + "loss": 6.641025543212891, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.039851363748311996, + "learning_rate": 0.0006, + "loss": 6.665952205657959, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.018417146056890488, + "learning_rate": 0.0006, + "loss": 6.657257556915283, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.02894527092576027, + "learning_rate": 0.0006, + "loss": 6.656288146972656, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.01910966821014881, + "learning_rate": 0.0006, + "loss": 6.688958168029785, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.01668412610888481, + "learning_rate": 0.0006, + "loss": 6.64876127243042, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02935275249183178, + "learning_rate": 0.0006, + "loss": 6.626096725463867, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.017997972667217255, + "learning_rate": 0.0006, + "loss": 6.640386581420898, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.01779189519584179, + "learning_rate": 0.0006, + "loss": 6.641165256500244, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.013901712372899055, + "learning_rate": 0.0006, + "loss": 6.651627063751221, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013211199082434177, + "learning_rate": 0.0006, + "loss": 6.630064964294434, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.02324107475578785, + "learning_rate": 0.0006, + "loss": 6.670443058013916, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.019452504813671112, + "learning_rate": 0.0006, + "loss": 6.651266098022461, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.017943572252988815, + "learning_rate": 0.0006, + "loss": 6.657943248748779, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.012136237695813179, + "learning_rate": 0.0006, + "loss": 6.640082359313965, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.011544011533260345, + "learning_rate": 0.0006, + "loss": 6.6186933517456055, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01569426618516445, + "learning_rate": 0.0006, + "loss": 6.640767574310303, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.027668394148349762, + "learning_rate": 0.0006, + "loss": 6.651412487030029, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.04073230177164078, + "learning_rate": 0.0006, + "loss": 6.636565208435059, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.04942672327160835, + "learning_rate": 0.0006, + "loss": 6.667994499206543, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.037568598985672, + "learning_rate": 0.0006, + "loss": 6.632008075714111, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.009880400262773037, + "learning_rate": 0.0006, + "loss": 6.6504058837890625, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.033384259790182114, + "learning_rate": 0.0006, + "loss": 6.63852596282959, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.035640884190797806, + "learning_rate": 0.0006, + "loss": 6.627986431121826, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.013637864030897617, + "learning_rate": 0.0006, + "loss": 6.624503135681152, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.025304675102233887, + "learning_rate": 0.0006, + "loss": 6.612077236175537, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.03377160802483559, + "learning_rate": 0.0006, + "loss": 6.608938217163086, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.019803153350949287, + "learning_rate": 0.0006, + "loss": 6.636343955993652, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.018928129225969315, + "learning_rate": 0.0006, + "loss": 6.636613845825195, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.023068996146321297, + "learning_rate": 0.0006, + "loss": 6.6258440017700195, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.016849957406520844, + "learning_rate": 0.0006, + "loss": 6.638483047485352, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.021430544555187225, + "learning_rate": 0.0006, + "loss": 6.619649410247803, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.024053368717432022, + "learning_rate": 0.0006, + "loss": 6.639214515686035, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.011926693841814995, + "learning_rate": 0.0006, + "loss": 6.604836463928223, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01961611397564411, + "learning_rate": 0.0006, + "loss": 6.624864101409912, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.027946067973971367, + "learning_rate": 0.0006, + "loss": 6.621284484863281, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.02108592912554741, + "learning_rate": 0.0006, + "loss": 6.5982465744018555, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.017272260040044785, + "learning_rate": 0.0006, + "loss": 6.626142501831055, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.015654582530260086, + "learning_rate": 0.0006, + "loss": 6.621735572814941, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.015797540545463562, + "learning_rate": 0.0006, + "loss": 6.6016411781311035, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.028488904237747192, + "learning_rate": 0.0006, + "loss": 6.598413467407227, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.03411734476685524, + "learning_rate": 0.0006, + "loss": 6.597321033477783, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.027188263833522797, + "learning_rate": 0.0006, + "loss": 6.630616188049316, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.03411437198519707, + "learning_rate": 0.0006, + "loss": 6.57647705078125, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.019907327368855476, + "learning_rate": 0.0006, + "loss": 6.579545974731445, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.02485930174589157, + "learning_rate": 0.0006, + "loss": 6.591073036193848, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.026106052100658417, + "learning_rate": 0.0006, + "loss": 6.580729961395264, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.029401391744613647, + "learning_rate": 0.0006, + "loss": 6.5460076332092285, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.032869379967451096, + "learning_rate": 0.0006, + "loss": 6.583278179168701, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 7.416288375854492, + "eval_runtime": 60.4177, + "eval_samples_per_second": 40.419, + "eval_steps_per_second": 1.274, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.04862433671951294, + "learning_rate": 0.0006, + "loss": 6.58473014831543, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.06636223942041397, + "learning_rate": 0.0006, + "loss": 6.621153354644775, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.05888264626264572, + "learning_rate": 0.0006, + "loss": 6.621196269989014, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.05331697314977646, + "learning_rate": 0.0006, + "loss": 6.600763320922852, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03947671875357628, + "learning_rate": 0.0006, + "loss": 6.600671768188477, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.04045343026518822, + "learning_rate": 0.0006, + "loss": 6.63529109954834, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.029766036197543144, + "learning_rate": 0.0006, + "loss": 6.612868309020996, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.03416327014565468, + "learning_rate": 0.0006, + "loss": 6.582240581512451, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03458515554666519, + "learning_rate": 0.0006, + "loss": 6.604381561279297, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.024099089205265045, + "learning_rate": 0.0006, + "loss": 6.597355365753174, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.02694438211619854, + "learning_rate": 0.0006, + "loss": 6.597512245178223, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024201856926083565, + "learning_rate": 0.0006, + "loss": 6.564685821533203, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.015959380194544792, + "learning_rate": 0.0006, + "loss": 6.582263946533203, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.025690171867609024, + "learning_rate": 0.0006, + "loss": 6.601057052612305, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.014497052878141403, + "learning_rate": 0.0006, + "loss": 6.584796905517578, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.019870297983288765, + "learning_rate": 0.0006, + "loss": 6.574010372161865, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.01807512901723385, + "learning_rate": 0.0006, + "loss": 6.558596134185791, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.014525193721055984, + "learning_rate": 0.0006, + "loss": 6.571803092956543, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.016113435849547386, + "learning_rate": 0.0006, + "loss": 6.568970680236816, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020724017173051834, + "learning_rate": 0.0006, + "loss": 6.5669403076171875, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.013963675126433372, + "learning_rate": 0.0006, + "loss": 6.586382865905762, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.01621115580201149, + "learning_rate": 0.0006, + "loss": 6.539546966552734, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.014925207942724228, + "learning_rate": 0.0006, + "loss": 6.531795501708984, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.015454350039362907, + "learning_rate": 0.0006, + "loss": 6.575699806213379, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.01623542234301567, + "learning_rate": 0.0006, + "loss": 6.582373142242432, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015452183783054352, + "learning_rate": 0.0006, + "loss": 6.576451301574707, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.021102432161569595, + "learning_rate": 0.0006, + "loss": 6.543055534362793, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.009827060624957085, + "learning_rate": 0.0006, + "loss": 6.526939392089844, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.013735136948525906, + "learning_rate": 0.0006, + "loss": 6.538825988769531, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02073509432375431, + "learning_rate": 0.0006, + "loss": 6.521315574645996, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.03110959194600582, + "learning_rate": 0.0006, + "loss": 6.561807632446289, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.045058365911245346, + "learning_rate": 0.0006, + "loss": 6.526515960693359, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.06359834969043732, + "learning_rate": 0.0006, + "loss": 6.554903030395508, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.06406767666339874, + "learning_rate": 0.0006, + "loss": 6.576272010803223, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.03403699770569801, + "learning_rate": 0.0006, + "loss": 6.5306243896484375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.036109618842601776, + "learning_rate": 0.0006, + "loss": 6.54176139831543, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.03478201478719711, + "learning_rate": 0.0006, + "loss": 6.530734062194824, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.03401544317603111, + "learning_rate": 0.0006, + "loss": 6.557794094085693, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.031629662960767746, + "learning_rate": 0.0006, + "loss": 6.564887046813965, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.02305024117231369, + "learning_rate": 0.0006, + "loss": 6.547412872314453, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030852800235152245, + "learning_rate": 0.0006, + "loss": 6.495370388031006, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.03368550166487694, + "learning_rate": 0.0006, + "loss": 6.495980262756348, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.020976359024643898, + "learning_rate": 0.0006, + "loss": 6.509410858154297, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.019652143120765686, + "learning_rate": 0.0006, + "loss": 6.493232727050781, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.024320121854543686, + "learning_rate": 0.0006, + "loss": 6.521531581878662, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.016611427068710327, + "learning_rate": 0.0006, + "loss": 6.507271766662598, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01635856367647648, + "learning_rate": 0.0006, + "loss": 6.492038726806641, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.017642095685005188, + "learning_rate": 0.0006, + "loss": 6.500349998474121, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.01679178513586521, + "learning_rate": 0.0006, + "loss": 6.500361442565918, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.011776307597756386, + "learning_rate": 0.0006, + "loss": 6.482490539550781, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.0176856629550457, + "learning_rate": 0.0006, + "loss": 6.48488712310791, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.015193155966699123, + "learning_rate": 0.0006, + "loss": 6.459545612335205, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.018623212352395058, + "learning_rate": 0.0006, + "loss": 6.478870391845703, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.012841006740927696, + "learning_rate": 0.0006, + "loss": 6.474418640136719, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.015540511347353458, + "learning_rate": 0.0006, + "loss": 6.4706525802612305, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.022695740684866905, + "learning_rate": 0.0006, + "loss": 6.455738067626953, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.026930129155516624, + "learning_rate": 0.0006, + "loss": 6.458016872406006, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.03555797412991524, + "learning_rate": 0.0006, + "loss": 6.480044364929199, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.05024682730436325, + "learning_rate": 0.0006, + "loss": 6.454564571380615, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.06616255640983582, + "learning_rate": 0.0006, + "loss": 6.433927536010742, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.059065770357847214, + "learning_rate": 0.0006, + "loss": 6.508362770080566, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.026689983904361725, + "learning_rate": 0.0006, + "loss": 6.464297294616699, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.04616398364305496, + "learning_rate": 0.0006, + "loss": 6.45448112487793, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.04980356618762016, + "learning_rate": 0.0006, + "loss": 6.4696574211120605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.0551983006298542, + "learning_rate": 0.0006, + "loss": 6.419000625610352, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.06988243758678436, + "learning_rate": 0.0006, + "loss": 6.442890167236328, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.06040226295590401, + "learning_rate": 0.0006, + "loss": 6.513258934020996, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04635024070739746, + "learning_rate": 0.0006, + "loss": 6.4710283279418945, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.04762702435255051, + "learning_rate": 0.0006, + "loss": 6.469327926635742, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.04857536032795906, + "learning_rate": 0.0006, + "loss": 6.426105499267578, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.04412693530321121, + "learning_rate": 0.0006, + "loss": 6.43179988861084, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026129672303795815, + "learning_rate": 0.0006, + "loss": 6.437918663024902, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 7.219249248504639, + "eval_runtime": 58.6837, + "eval_samples_per_second": 41.613, + "eval_steps_per_second": 1.312, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03329874947667122, + "learning_rate": 0.0006, + "loss": 6.438179969787598, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03723187372088432, + "learning_rate": 0.0006, + "loss": 6.394242286682129, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.014625726267695427, + "learning_rate": 0.0006, + "loss": 6.412118911743164, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.02917475439608097, + "learning_rate": 0.0006, + "loss": 6.453164577484131, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01811208389699459, + "learning_rate": 0.0006, + "loss": 6.396003246307373, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.027235351502895355, + "learning_rate": 0.0006, + "loss": 6.402976036071777, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.02835794910788536, + "learning_rate": 0.0006, + "loss": 6.380568504333496, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.020868154242634773, + "learning_rate": 0.0006, + "loss": 6.38240909576416, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.02113114856183529, + "learning_rate": 0.0006, + "loss": 6.377391815185547, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.018894119188189507, + "learning_rate": 0.0006, + "loss": 6.354569435119629, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.01820225641131401, + "learning_rate": 0.0006, + "loss": 6.392882347106934, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.017510728910565376, + "learning_rate": 0.0006, + "loss": 6.364162445068359, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.017893686890602112, + "learning_rate": 0.0006, + "loss": 6.373456954956055, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.013382102362811565, + "learning_rate": 0.0006, + "loss": 6.3714399337768555, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.022540833801031113, + "learning_rate": 0.0006, + "loss": 6.348170757293701, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.02110104262828827, + "learning_rate": 0.0006, + "loss": 6.413206100463867, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.014498349279165268, + "learning_rate": 0.0006, + "loss": 6.361384391784668, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.01819637045264244, + "learning_rate": 0.0006, + "loss": 6.315275192260742, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.0195972491055727, + "learning_rate": 0.0006, + "loss": 6.321994781494141, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.018822453916072845, + "learning_rate": 0.0006, + "loss": 6.345219612121582, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.019247712567448616, + "learning_rate": 0.0006, + "loss": 6.348073959350586, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.03175375238060951, + "learning_rate": 0.0006, + "loss": 6.300344944000244, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.06326444447040558, + "learning_rate": 0.0006, + "loss": 6.392579078674316, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.09667232632637024, + "learning_rate": 0.0006, + "loss": 6.36024284362793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.05774446204304695, + "learning_rate": 0.0006, + "loss": 6.365338325500488, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.08793356269598007, + "learning_rate": 0.0006, + "loss": 6.43284273147583, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.0640595331788063, + "learning_rate": 0.0006, + "loss": 6.342428207397461, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.0447838194668293, + "learning_rate": 0.0006, + "loss": 6.367971897125244, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.04768967255949974, + "learning_rate": 0.0006, + "loss": 6.35027551651001, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.027981851249933243, + "learning_rate": 0.0006, + "loss": 6.33723258972168, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.03479623422026634, + "learning_rate": 0.0006, + "loss": 6.36590051651001, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.020846687257289886, + "learning_rate": 0.0006, + "loss": 6.315143585205078, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.0312928780913353, + "learning_rate": 0.0006, + "loss": 6.340150833129883, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.020593881607055664, + "learning_rate": 0.0006, + "loss": 6.348175048828125, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.022010236978530884, + "learning_rate": 0.0006, + "loss": 6.318607330322266, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.018427059054374695, + "learning_rate": 0.0006, + "loss": 6.331822395324707, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.019964968785643578, + "learning_rate": 0.0006, + "loss": 6.287320137023926, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.02130270190536976, + "learning_rate": 0.0006, + "loss": 6.353344917297363, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.01616514101624489, + "learning_rate": 0.0006, + "loss": 6.314052581787109, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.016639474779367447, + "learning_rate": 0.0006, + "loss": 6.320351600646973, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.017620692029595375, + "learning_rate": 0.0006, + "loss": 6.311459064483643, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.013542844913899899, + "learning_rate": 0.0006, + "loss": 6.324368476867676, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01308678649365902, + "learning_rate": 0.0006, + "loss": 6.307419776916504, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.014672392047941685, + "learning_rate": 0.0006, + "loss": 6.2785420417785645, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01399585697799921, + "learning_rate": 0.0006, + "loss": 6.258619785308838, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.01330904383212328, + "learning_rate": 0.0006, + "loss": 6.299595832824707, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016613377258181572, + "learning_rate": 0.0006, + "loss": 6.2900309562683105, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.015076331794261932, + "learning_rate": 0.0006, + "loss": 6.231028079986572, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.01589728519320488, + "learning_rate": 0.0006, + "loss": 6.257922172546387, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.030972328037023544, + "learning_rate": 0.0006, + "loss": 6.285238265991211, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.05048783868551254, + "learning_rate": 0.0006, + "loss": 6.287471771240234, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.07390714436769485, + "learning_rate": 0.0006, + "loss": 6.284293174743652, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.07166943699121475, + "learning_rate": 0.0006, + "loss": 6.290704727172852, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.028792452067136765, + "learning_rate": 0.0006, + "loss": 6.243415832519531, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.04822041466832161, + "learning_rate": 0.0006, + "loss": 6.278229713439941, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.036688923835754395, + "learning_rate": 0.0006, + "loss": 6.275590896606445, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.03679114952683449, + "learning_rate": 0.0006, + "loss": 6.236509323120117, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.023249102756381035, + "learning_rate": 0.0006, + "loss": 6.230788230895996, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.02682286500930786, + "learning_rate": 0.0006, + "loss": 6.278890609741211, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.02202044241130352, + "learning_rate": 0.0006, + "loss": 6.234531402587891, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.02545459009706974, + "learning_rate": 0.0006, + "loss": 6.271790981292725, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.02012932486832142, + "learning_rate": 0.0006, + "loss": 6.247462749481201, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.02216474711894989, + "learning_rate": 0.0006, + "loss": 6.26396369934082, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.016078704968094826, + "learning_rate": 0.0006, + "loss": 6.228609085083008, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.023021478205919266, + "learning_rate": 0.0006, + "loss": 6.197244167327881, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.024766333401203156, + "learning_rate": 0.0006, + "loss": 6.181090354919434, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.025581466034054756, + "learning_rate": 0.0006, + "loss": 6.201817512512207, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.016964981332421303, + "learning_rate": 0.0006, + "loss": 6.201486587524414, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.023957721889019012, + "learning_rate": 0.0006, + "loss": 6.185701847076416, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.029775245115160942, + "learning_rate": 0.0006, + "loss": 6.242060661315918, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04007191210985184, + "learning_rate": 0.0006, + "loss": 6.19931697845459, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04882276430726051, + "learning_rate": 0.0006, + "loss": 6.1880598068237305, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 7.018043518066406, + "eval_runtime": 59.8815, + "eval_samples_per_second": 40.781, + "eval_steps_per_second": 1.286, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.07144106924533844, + "learning_rate": 0.0006, + "loss": 6.172799110412598, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.06712841987609863, + "learning_rate": 0.0006, + "loss": 6.228682518005371, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.04405239224433899, + "learning_rate": 0.0006, + "loss": 6.1827898025512695, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.03418995440006256, + "learning_rate": 0.0006, + "loss": 6.200876235961914, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.04002382606267929, + "learning_rate": 0.0006, + "loss": 6.1799211502075195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.029748762026429176, + "learning_rate": 0.0006, + "loss": 6.125422477722168, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03408357501029968, + "learning_rate": 0.0006, + "loss": 6.203458309173584, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.02227279357612133, + "learning_rate": 0.0006, + "loss": 6.135439872741699, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02727862447500229, + "learning_rate": 0.0006, + "loss": 6.202380180358887, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.026019075885415077, + "learning_rate": 0.0006, + "loss": 6.201343536376953, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.027227282524108887, + "learning_rate": 0.0006, + "loss": 6.164059638977051, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.027649980038404465, + "learning_rate": 0.0006, + "loss": 6.135448932647705, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02168039232492447, + "learning_rate": 0.0006, + "loss": 6.154831886291504, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02101467363536358, + "learning_rate": 0.0006, + "loss": 6.128695487976074, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.018344245851039886, + "learning_rate": 0.0006, + "loss": 6.136524200439453, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020323453471064568, + "learning_rate": 0.0006, + "loss": 6.1206488609313965, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.01750849559903145, + "learning_rate": 0.0006, + "loss": 6.128170967102051, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017778221517801285, + "learning_rate": 0.0006, + "loss": 6.150458812713623, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.01863938942551613, + "learning_rate": 0.0006, + "loss": 6.100106716156006, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.020196080207824707, + "learning_rate": 0.0006, + "loss": 6.149346351623535, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.03203890845179558, + "learning_rate": 0.0006, + "loss": 6.124648094177246, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.05325170233845711, + "learning_rate": 0.0006, + "loss": 6.115618705749512, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.05888345465064049, + "learning_rate": 0.0006, + "loss": 6.165221214294434, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.026873115450143814, + "learning_rate": 0.0006, + "loss": 6.039590835571289, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.06304649263620377, + "learning_rate": 0.0006, + "loss": 6.131900787353516, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.08567557483911514, + "learning_rate": 0.0006, + "loss": 6.218114852905273, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.08868218213319778, + "learning_rate": 0.0006, + "loss": 6.216706275939941, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.03955601155757904, + "learning_rate": 0.0006, + "loss": 6.1566925048828125, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.04979498311877251, + "learning_rate": 0.0006, + "loss": 6.19327449798584, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.03253026679158211, + "learning_rate": 0.0006, + "loss": 6.133993148803711, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03016652725636959, + "learning_rate": 0.0006, + "loss": 6.174962997436523, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.03159862384200096, + "learning_rate": 0.0006, + "loss": 6.082124710083008, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.027673959732055664, + "learning_rate": 0.0006, + "loss": 6.111873149871826, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.022789714857935905, + "learning_rate": 0.0006, + "loss": 6.129437446594238, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.030061012133955956, + "learning_rate": 0.0006, + "loss": 6.126370429992676, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.019007068127393723, + "learning_rate": 0.0006, + "loss": 6.07887077331543, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.01897653006017208, + "learning_rate": 0.0006, + "loss": 6.096936225891113, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.024659698829054832, + "learning_rate": 0.0006, + "loss": 6.0933380126953125, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.021924549713730812, + "learning_rate": 0.0006, + "loss": 6.105311870574951, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.018817558884620667, + "learning_rate": 0.0006, + "loss": 6.058895111083984, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.02477840520441532, + "learning_rate": 0.0006, + "loss": 6.038928031921387, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.018968632444739342, + "learning_rate": 0.0006, + "loss": 6.1368937492370605, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.02409801445901394, + "learning_rate": 0.0006, + "loss": 6.110324859619141, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.017595326527953148, + "learning_rate": 0.0006, + "loss": 6.074368476867676, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.01875406503677368, + "learning_rate": 0.0006, + "loss": 6.121023654937744, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.018781628459692, + "learning_rate": 0.0006, + "loss": 6.064914703369141, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.017003126442432404, + "learning_rate": 0.0006, + "loss": 6.039608001708984, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.017665037885308266, + "learning_rate": 0.0006, + "loss": 6.053671836853027, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.02066265605390072, + "learning_rate": 0.0006, + "loss": 6.090536117553711, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02888769656419754, + "learning_rate": 0.0006, + "loss": 5.9858903884887695, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.03744485229253769, + "learning_rate": 0.0006, + "loss": 6.086264610290527, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04719431325793266, + "learning_rate": 0.0006, + "loss": 6.09747314453125, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.044963907450437546, + "learning_rate": 0.0006, + "loss": 6.091272354125977, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03810194507241249, + "learning_rate": 0.0006, + "loss": 6.084904193878174, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.04626769199967384, + "learning_rate": 0.0006, + "loss": 6.017384052276611, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.03459552302956581, + "learning_rate": 0.0006, + "loss": 6.111314296722412, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.027385666966438293, + "learning_rate": 0.0006, + "loss": 6.002509117126465, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.03417652100324631, + "learning_rate": 0.0006, + "loss": 6.0421061515808105, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.04415450245141983, + "learning_rate": 0.0006, + "loss": 6.035305023193359, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.03259635716676712, + "learning_rate": 0.0006, + "loss": 6.025840759277344, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.024941129609942436, + "learning_rate": 0.0006, + "loss": 6.043416976928711, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.02434500865638256, + "learning_rate": 0.0006, + "loss": 6.002052307128906, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.01876584254205227, + "learning_rate": 0.0006, + "loss": 6.045035362243652, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.02254367619752884, + "learning_rate": 0.0006, + "loss": 5.972503662109375, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.024504603818058968, + "learning_rate": 0.0006, + "loss": 5.992846488952637, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.02317827194929123, + "learning_rate": 0.0006, + "loss": 6.007697582244873, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02225821278989315, + "learning_rate": 0.0006, + "loss": 6.01463508605957, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.01811717264354229, + "learning_rate": 0.0006, + "loss": 5.995022773742676, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.019589431583881378, + "learning_rate": 0.0006, + "loss": 5.990633964538574, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015422016382217407, + "learning_rate": 0.0006, + "loss": 6.011263847351074, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.9954938888549805, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.022172044962644577, + "learning_rate": 0.0006, + "loss": 5.95679235458374, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 6.75809383392334, + "eval_runtime": 72.8926, + "eval_samples_per_second": 33.501, + "eval_steps_per_second": 1.056, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.029082221910357475, + "learning_rate": 0.0006, + "loss": 6.018675804138184, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.03959495946764946, + "learning_rate": 0.0006, + "loss": 5.93310546875, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.043756261467933655, + "learning_rate": 0.0006, + "loss": 5.965944290161133, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.04996631667017937, + "learning_rate": 0.0006, + "loss": 6.0381059646606445, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.05721999332308769, + "learning_rate": 0.0006, + "loss": 5.968381881713867, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.07594525814056396, + "learning_rate": 0.0006, + "loss": 6.027225494384766, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.06861685961484909, + "learning_rate": 0.0006, + "loss": 6.035175323486328, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04156472906470299, + "learning_rate": 0.0006, + "loss": 5.9500017166137695, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.045399412512779236, + "learning_rate": 0.0006, + "loss": 5.9953389167785645, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.04399585723876953, + "learning_rate": 0.0006, + "loss": 5.986051559448242, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05764509737491608, + "learning_rate": 0.0006, + "loss": 6.005392074584961, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.060871850699186325, + "learning_rate": 0.0006, + "loss": 6.01304817199707, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.03381241112947464, + "learning_rate": 0.0006, + "loss": 5.977019309997559, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.03384104371070862, + "learning_rate": 0.0006, + "loss": 5.988002777099609, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.03120286948978901, + "learning_rate": 0.0006, + "loss": 5.968227386474609, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.026532653719186783, + "learning_rate": 0.0006, + "loss": 5.936810493469238, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.024956922978162766, + "learning_rate": 0.0006, + "loss": 5.9348015785217285, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.025926144793629646, + "learning_rate": 0.0006, + "loss": 5.952630519866943, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.02893793396651745, + "learning_rate": 0.0006, + "loss": 5.9569854736328125, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.0320001058280468, + "learning_rate": 0.0006, + "loss": 5.917050838470459, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.03925463184714317, + "learning_rate": 0.0006, + "loss": 5.9925127029418945, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.05096640810370445, + "learning_rate": 0.0006, + "loss": 5.926358222961426, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.041637953370809555, + "learning_rate": 0.0006, + "loss": 5.954983711242676, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03138125315308571, + "learning_rate": 0.0006, + "loss": 5.893692970275879, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.029299303889274597, + "learning_rate": 0.0006, + "loss": 5.94670295715332, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.023969192057847977, + "learning_rate": 0.0006, + "loss": 5.925658702850342, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.025406932458281517, + "learning_rate": 0.0006, + "loss": 5.966875076293945, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.024309059605002403, + "learning_rate": 0.0006, + "loss": 5.932581424713135, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.017237074673175812, + "learning_rate": 0.0006, + "loss": 5.898823261260986, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.025594931095838547, + "learning_rate": 0.0006, + "loss": 5.9285993576049805, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.025184160098433495, + "learning_rate": 0.0006, + "loss": 5.915966510772705, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019004398956894875, + "learning_rate": 0.0006, + "loss": 5.893395900726318, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.021802261471748352, + "learning_rate": 0.0006, + "loss": 5.953311443328857, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02202655002474785, + "learning_rate": 0.0006, + "loss": 5.953000545501709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.02067241445183754, + "learning_rate": 0.0006, + "loss": 5.869247913360596, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.015430927276611328, + "learning_rate": 0.0006, + "loss": 5.855883598327637, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.019917506724596024, + "learning_rate": 0.0006, + "loss": 5.9284539222717285, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.024291323497891426, + "learning_rate": 0.0006, + "loss": 5.973604202270508, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.022133296355605125, + "learning_rate": 0.0006, + "loss": 5.85693359375, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0254366435110569, + "learning_rate": 0.0006, + "loss": 5.939558982849121, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.030134694650769234, + "learning_rate": 0.0006, + "loss": 5.892992973327637, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.03845192864537239, + "learning_rate": 0.0006, + "loss": 5.89303731918335, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.042409464716911316, + "learning_rate": 0.0006, + "loss": 5.914068222045898, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.0454271174967289, + "learning_rate": 0.0006, + "loss": 5.853579044342041, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.04934931918978691, + "learning_rate": 0.0006, + "loss": 5.881861686706543, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.0519695058465004, + "learning_rate": 0.0006, + "loss": 5.872620582580566, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.06172563508152962, + "learning_rate": 0.0006, + "loss": 5.938283443450928, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.08416368812322617, + "learning_rate": 0.0006, + "loss": 5.943403720855713, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.07311230152845383, + "learning_rate": 0.0006, + "loss": 5.954029083251953, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.09851036965847015, + "learning_rate": 0.0006, + "loss": 5.992746829986572, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.07521767914295197, + "learning_rate": 0.0006, + "loss": 5.956884384155273, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.05720978602766991, + "learning_rate": 0.0006, + "loss": 5.911205291748047, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.05360936373472214, + "learning_rate": 0.0006, + "loss": 5.9734787940979, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.035826463252305984, + "learning_rate": 0.0006, + "loss": 5.95751953125, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.030137093737721443, + "learning_rate": 0.0006, + "loss": 5.91911506652832, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030039150267839432, + "learning_rate": 0.0006, + "loss": 5.910290718078613, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.03349437564611435, + "learning_rate": 0.0006, + "loss": 5.904623985290527, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02211189828813076, + "learning_rate": 0.0006, + "loss": 5.909668445587158, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.022601254284381866, + "learning_rate": 0.0006, + "loss": 5.942885875701904, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.02197270467877388, + "learning_rate": 0.0006, + "loss": 5.867103576660156, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.022765018045902252, + "learning_rate": 0.0006, + "loss": 5.894528865814209, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021181151270866394, + "learning_rate": 0.0006, + "loss": 5.896910667419434, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.01878993771970272, + "learning_rate": 0.0006, + "loss": 5.860502243041992, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.014679803512990475, + "learning_rate": 0.0006, + "loss": 5.8725266456604, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.017919695004820824, + "learning_rate": 0.0006, + "loss": 5.88902473449707, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.01437009871006012, + "learning_rate": 0.0006, + "loss": 5.868186950683594, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.014293290674686432, + "learning_rate": 0.0006, + "loss": 5.833364963531494, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.015440084971487522, + "learning_rate": 0.0006, + "loss": 5.860260009765625, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.014292514882981777, + "learning_rate": 0.0006, + "loss": 5.844939231872559, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.015702251344919205, + "learning_rate": 0.0006, + "loss": 5.825929164886475, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.013372798450291157, + "learning_rate": 0.0006, + "loss": 5.83062744140625, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.013794616796076298, + "learning_rate": 0.0006, + "loss": 5.7897820472717285, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 6.600830078125, + "eval_runtime": 71.4316, + "eval_samples_per_second": 34.187, + "eval_steps_per_second": 1.078, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.01262663770467043, + "learning_rate": 0.0006, + "loss": 5.802824020385742, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.016375157982110977, + "learning_rate": 0.0006, + "loss": 5.8018317222595215, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.010541851632297039, + "learning_rate": 0.0006, + "loss": 5.848208427429199, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.011726257391273975, + "learning_rate": 0.0006, + "loss": 5.771823883056641, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01124380249530077, + "learning_rate": 0.0006, + "loss": 5.773855209350586, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.010801885277032852, + "learning_rate": 0.0006, + "loss": 5.847390651702881, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.013018375262618065, + "learning_rate": 0.0006, + "loss": 5.786786079406738, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.014233003370463848, + "learning_rate": 0.0006, + "loss": 5.7913713455200195, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.01465831883251667, + "learning_rate": 0.0006, + "loss": 5.821170330047607, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.017383430153131485, + "learning_rate": 0.0006, + "loss": 5.811469078063965, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.020328007638454437, + "learning_rate": 0.0006, + "loss": 5.718976020812988, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.01776740700006485, + "learning_rate": 0.0006, + "loss": 5.80593729019165, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.017713800072669983, + "learning_rate": 0.0006, + "loss": 5.744930267333984, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.028774067759513855, + "learning_rate": 0.0006, + "loss": 5.85262393951416, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.05153049901127815, + "learning_rate": 0.0006, + "loss": 5.739948272705078, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.06882233172655106, + "learning_rate": 0.0006, + "loss": 5.844270706176758, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.0420759879052639, + "learning_rate": 0.0006, + "loss": 5.79060173034668, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.05598323419690132, + "learning_rate": 0.0006, + "loss": 5.833735466003418, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.05011717975139618, + "learning_rate": 0.0006, + "loss": 5.840887069702148, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.040782492607831955, + "learning_rate": 0.0006, + "loss": 5.7905449867248535, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.030804185196757317, + "learning_rate": 0.0006, + "loss": 5.802691459655762, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.032001759856939316, + "learning_rate": 0.0006, + "loss": 5.726612091064453, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.0295492485165596, + "learning_rate": 0.0006, + "loss": 5.798826694488525, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.034406546503305435, + "learning_rate": 0.0006, + "loss": 5.720219135284424, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.02627669647336006, + "learning_rate": 0.0006, + "loss": 5.723702430725098, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.0282363910228014, + "learning_rate": 0.0006, + "loss": 5.818911552429199, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.017894433811306953, + "learning_rate": 0.0006, + "loss": 5.772168159484863, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.019017303362488747, + "learning_rate": 0.0006, + "loss": 5.774587631225586, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.022159283980727196, + "learning_rate": 0.0006, + "loss": 5.812951564788818, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02198116108775139, + "learning_rate": 0.0006, + "loss": 5.808734893798828, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.017741188406944275, + "learning_rate": 0.0006, + "loss": 5.776142120361328, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.017197400331497192, + "learning_rate": 0.0006, + "loss": 5.7471208572387695, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.01752566732466221, + "learning_rate": 0.0006, + "loss": 5.780961990356445, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.017961280420422554, + "learning_rate": 0.0006, + "loss": 5.713686943054199, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.02227521874010563, + "learning_rate": 0.0006, + "loss": 5.717774868011475, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.024075504392385483, + "learning_rate": 0.0006, + "loss": 5.715015411376953, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.029195033013820648, + "learning_rate": 0.0006, + "loss": 5.659659385681152, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.02959120273590088, + "learning_rate": 0.0006, + "loss": 5.78981351852417, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.021854368969798088, + "learning_rate": 0.0006, + "loss": 5.750844955444336, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017120081931352615, + "learning_rate": 0.0006, + "loss": 5.672736167907715, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.020712822675704956, + "learning_rate": 0.0006, + "loss": 5.708046913146973, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.019186170771718025, + "learning_rate": 0.0006, + "loss": 5.71742057800293, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.01985814794898033, + "learning_rate": 0.0006, + "loss": 5.735311508178711, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.01965234987437725, + "learning_rate": 0.0006, + "loss": 5.728727340698242, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.019827688112854958, + "learning_rate": 0.0006, + "loss": 5.743068695068359, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.025500580668449402, + "learning_rate": 0.0006, + "loss": 5.76585578918457, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.03701791539788246, + "learning_rate": 0.0006, + "loss": 5.728367805480957, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.04173659533262253, + "learning_rate": 0.0006, + "loss": 5.683089256286621, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.031390558928251266, + "learning_rate": 0.0006, + "loss": 5.740921974182129, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.03970158472657204, + "learning_rate": 0.0006, + "loss": 5.730632781982422, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.03865141049027443, + "learning_rate": 0.0006, + "loss": 5.686657905578613, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.04856107011437416, + "learning_rate": 0.0006, + "loss": 5.657513618469238, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.04016133397817612, + "learning_rate": 0.0006, + "loss": 5.754552841186523, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.03961458429694176, + "learning_rate": 0.0006, + "loss": 5.672088623046875, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.03403536602854729, + "learning_rate": 0.0006, + "loss": 5.736673355102539, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03766333684325218, + "learning_rate": 0.0006, + "loss": 5.716219902038574, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.04085501655936241, + "learning_rate": 0.0006, + "loss": 5.678958415985107, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.036888182163238525, + "learning_rate": 0.0006, + "loss": 5.6990461349487305, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.0492791011929512, + "learning_rate": 0.0006, + "loss": 5.78053092956543, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.05572429671883583, + "learning_rate": 0.0006, + "loss": 5.7543535232543945, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.047000352293252945, + "learning_rate": 0.0006, + "loss": 5.7174272537231445, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.058798179030418396, + "learning_rate": 0.0006, + "loss": 5.793713569641113, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.046992287039756775, + "learning_rate": 0.0006, + "loss": 5.784457683563232, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.04382370039820671, + "learning_rate": 0.0006, + "loss": 5.793731212615967, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.040616173297166824, + "learning_rate": 0.0006, + "loss": 5.69599723815918, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03446170687675476, + "learning_rate": 0.0006, + "loss": 5.694674968719482, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029936550185084343, + "learning_rate": 0.0006, + "loss": 5.6704487800598145, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02224753610789776, + "learning_rate": 0.0006, + "loss": 5.728894233703613, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.024346651509404182, + "learning_rate": 0.0006, + "loss": 5.772412300109863, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.022997068241238594, + "learning_rate": 0.0006, + "loss": 5.645310401916504, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.024918686598539352, + "learning_rate": 0.0006, + "loss": 5.671009063720703, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.020067067816853523, + "learning_rate": 0.0006, + "loss": 5.637441635131836, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 6.387807846069336, + "eval_runtime": 72.3372, + "eval_samples_per_second": 33.759, + "eval_steps_per_second": 1.064, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02015046589076519, + "learning_rate": 0.0006, + "loss": 5.598219871520996, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.017398346215486526, + "learning_rate": 0.0006, + "loss": 5.66834831237793, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.016871167346835136, + "learning_rate": 0.0006, + "loss": 5.673916816711426, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.01695060543715954, + "learning_rate": 0.0006, + "loss": 5.646677017211914, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.019298404455184937, + "learning_rate": 0.0006, + "loss": 5.663064479827881, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018699361011385918, + "learning_rate": 0.0006, + "loss": 5.615242958068848, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.019567854702472687, + "learning_rate": 0.0006, + "loss": 5.5521240234375, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.018347466364502907, + "learning_rate": 0.0006, + "loss": 5.5929107666015625, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.020350053906440735, + "learning_rate": 0.0006, + "loss": 5.6380109786987305, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.016790535300970078, + "learning_rate": 0.0006, + "loss": 5.62994384765625, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.01413649506866932, + "learning_rate": 0.0006, + "loss": 5.6301727294921875, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.013762339949607849, + "learning_rate": 0.0006, + "loss": 5.584138870239258, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.014790690504014492, + "learning_rate": 0.0006, + "loss": 5.666123390197754, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.015471262857317924, + "learning_rate": 0.0006, + "loss": 5.587042808532715, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.015804501250386238, + "learning_rate": 0.0006, + "loss": 5.627497673034668, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.017627960070967674, + "learning_rate": 0.0006, + "loss": 5.556890487670898, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.02074616774916649, + "learning_rate": 0.0006, + "loss": 5.586511135101318, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.026389935985207558, + "learning_rate": 0.0006, + "loss": 5.598520278930664, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03736288473010063, + "learning_rate": 0.0006, + "loss": 5.575788497924805, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.03523678705096245, + "learning_rate": 0.0006, + "loss": 5.586930751800537, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.024266984313726425, + "learning_rate": 0.0006, + "loss": 5.596694469451904, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.033640433102846146, + "learning_rate": 0.0006, + "loss": 5.536456108093262, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.04689764603972435, + "learning_rate": 0.0006, + "loss": 5.61761999130249, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.04910685867071152, + "learning_rate": 0.0006, + "loss": 5.6298418045043945, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.03152550011873245, + "learning_rate": 0.0006, + "loss": 5.639697074890137, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.037753574550151825, + "learning_rate": 0.0006, + "loss": 5.697381019592285, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.04877670109272003, + "learning_rate": 0.0006, + "loss": 5.605184555053711, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.04228494316339493, + "learning_rate": 0.0006, + "loss": 5.624238967895508, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.03575790300965309, + "learning_rate": 0.0006, + "loss": 5.6037983894348145, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.036425549536943436, + "learning_rate": 0.0006, + "loss": 5.62905216217041, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.04001355543732643, + "learning_rate": 0.0006, + "loss": 5.581656455993652, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.05040939524769783, + "learning_rate": 0.0006, + "loss": 5.621358871459961, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.04140249267220497, + "learning_rate": 0.0006, + "loss": 5.668188095092773, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.033116403967142105, + "learning_rate": 0.0006, + "loss": 5.580477714538574, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.03864264488220215, + "learning_rate": 0.0006, + "loss": 5.539153099060059, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.03026588261127472, + "learning_rate": 0.0006, + "loss": 5.619814872741699, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.02381783165037632, + "learning_rate": 0.0006, + "loss": 5.64422607421875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02427830547094345, + "learning_rate": 0.0006, + "loss": 5.659873008728027, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.021709900349378586, + "learning_rate": 0.0006, + "loss": 5.590245723724365, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.022812914103269577, + "learning_rate": 0.0006, + "loss": 5.610434532165527, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.019250890240073204, + "learning_rate": 0.0006, + "loss": 5.588078498840332, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.021995501592755318, + "learning_rate": 0.0006, + "loss": 5.541754245758057, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01477806642651558, + "learning_rate": 0.0006, + "loss": 5.549173355102539, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.016680821776390076, + "learning_rate": 0.0006, + "loss": 5.58221435546875, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013848318718373775, + "learning_rate": 0.0006, + "loss": 5.5505876541137695, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.016387000679969788, + "learning_rate": 0.0006, + "loss": 5.570460319519043, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.016496233642101288, + "learning_rate": 0.0006, + "loss": 5.513058662414551, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01675555482506752, + "learning_rate": 0.0006, + "loss": 5.559998989105225, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.016939565539360046, + "learning_rate": 0.0006, + "loss": 5.56727409362793, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.01679506152868271, + "learning_rate": 0.0006, + "loss": 5.603643417358398, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01656349189579487, + "learning_rate": 0.0006, + "loss": 5.54292631149292, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.01975332945585251, + "learning_rate": 0.0006, + "loss": 5.5225677490234375, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.02329389750957489, + "learning_rate": 0.0006, + "loss": 5.537827491760254, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.026616891846060753, + "learning_rate": 0.0006, + "loss": 5.500011920928955, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.025511575862765312, + "learning_rate": 0.0006, + "loss": 5.5316619873046875, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.025009512901306152, + "learning_rate": 0.0006, + "loss": 5.468493938446045, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02946239896118641, + "learning_rate": 0.0006, + "loss": 5.5116753578186035, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.036607250571250916, + "learning_rate": 0.0006, + "loss": 5.518636226654053, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.04347814992070198, + "learning_rate": 0.0006, + "loss": 5.572815895080566, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.034463658928871155, + "learning_rate": 0.0006, + "loss": 5.520634651184082, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.03486703336238861, + "learning_rate": 0.0006, + "loss": 5.545441150665283, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.04476739838719368, + "learning_rate": 0.0006, + "loss": 5.554450035095215, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.048745233565568924, + "learning_rate": 0.0006, + "loss": 5.554961204528809, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.04636877775192261, + "learning_rate": 0.0006, + "loss": 5.574655055999756, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.032686877995729446, + "learning_rate": 0.0006, + "loss": 5.5734639167785645, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.03075302764773369, + "learning_rate": 0.0006, + "loss": 5.487662315368652, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.028860175982117653, + "learning_rate": 0.0006, + "loss": 5.519798755645752, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.028599515557289124, + "learning_rate": 0.0006, + "loss": 5.558349609375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.02384260669350624, + "learning_rate": 0.0006, + "loss": 5.520661354064941, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.023180771619081497, + "learning_rate": 0.0006, + "loss": 5.496896743774414, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.021167006343603134, + "learning_rate": 0.0006, + "loss": 5.517192363739014, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02069013938307762, + "learning_rate": 0.0006, + "loss": 5.447656154632568, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 6.161121845245361, + "eval_runtime": 71.5972, + "eval_samples_per_second": 34.107, + "eval_steps_per_second": 1.075, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.022983292117714882, + "learning_rate": 0.0006, + "loss": 5.441346645355225, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.022833511233329773, + "learning_rate": 0.0006, + "loss": 5.45123815536499, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.020015383139252663, + "learning_rate": 0.0006, + "loss": 5.430389404296875, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.016957225278019905, + "learning_rate": 0.0006, + "loss": 5.473883628845215, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.02090892568230629, + "learning_rate": 0.0006, + "loss": 5.467441558837891, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.022806301712989807, + "learning_rate": 0.0006, + "loss": 5.5453081130981445, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.022816946730017662, + "learning_rate": 0.0006, + "loss": 5.426213264465332, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.020761270076036453, + "learning_rate": 0.0006, + "loss": 5.464138507843018, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01982618309557438, + "learning_rate": 0.0006, + "loss": 5.428033828735352, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.021953724324703217, + "learning_rate": 0.0006, + "loss": 5.456646919250488, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01871841959655285, + "learning_rate": 0.0006, + "loss": 5.508403778076172, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.01810353808104992, + "learning_rate": 0.0006, + "loss": 5.442957878112793, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.017609186470508575, + "learning_rate": 0.0006, + "loss": 5.4774699211120605, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.01931103691458702, + "learning_rate": 0.0006, + "loss": 5.429748058319092, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.019401391968131065, + "learning_rate": 0.0006, + "loss": 5.466834545135498, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.018862580880522728, + "learning_rate": 0.0006, + "loss": 5.431701183319092, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.01796964928507805, + "learning_rate": 0.0006, + "loss": 5.388613700866699, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.021166296675801277, + "learning_rate": 0.0006, + "loss": 5.454118251800537, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.028568729758262634, + "learning_rate": 0.0006, + "loss": 5.510260581970215, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.03902891278266907, + "learning_rate": 0.0006, + "loss": 5.386396408081055, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.039114370942115784, + "learning_rate": 0.0006, + "loss": 5.457526206970215, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.03560924902558327, + "learning_rate": 0.0006, + "loss": 5.4149980545043945, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.03745810315012932, + "learning_rate": 0.0006, + "loss": 5.464084625244141, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.03383171558380127, + "learning_rate": 0.0006, + "loss": 5.45137882232666, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.027873128652572632, + "learning_rate": 0.0006, + "loss": 5.485784530639648, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.02697703056037426, + "learning_rate": 0.0006, + "loss": 5.459152698516846, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.02646288089454174, + "learning_rate": 0.0006, + "loss": 5.433281421661377, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.02655179798603058, + "learning_rate": 0.0006, + "loss": 5.456627368927002, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02633797563612461, + "learning_rate": 0.0006, + "loss": 5.484803676605225, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.031187085434794426, + "learning_rate": 0.0006, + "loss": 5.362293243408203, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02971349097788334, + "learning_rate": 0.0006, + "loss": 5.4369611740112305, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.02996021695435047, + "learning_rate": 0.0006, + "loss": 5.419068813323975, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.032519739121198654, + "learning_rate": 0.0006, + "loss": 5.422189712524414, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.03403490409255028, + "learning_rate": 0.0006, + "loss": 5.470074653625488, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.036989886313676834, + "learning_rate": 0.0006, + "loss": 5.460174560546875, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.03427419438958168, + "learning_rate": 0.0006, + "loss": 5.495920658111572, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.02519620768725872, + "learning_rate": 0.0006, + "loss": 5.474429130554199, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.02473096363246441, + "learning_rate": 0.0006, + "loss": 5.395847320556641, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.02792714536190033, + "learning_rate": 0.0006, + "loss": 5.376262664794922, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.024172596633434296, + "learning_rate": 0.0006, + "loss": 5.459731101989746, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02304821088910103, + "learning_rate": 0.0006, + "loss": 5.479475498199463, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.02342340536415577, + "learning_rate": 0.0006, + "loss": 5.387955665588379, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021962853148579597, + "learning_rate": 0.0006, + "loss": 5.385697364807129, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.019193042069673538, + "learning_rate": 0.0006, + "loss": 5.431601047515869, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.021378040313720703, + "learning_rate": 0.0006, + "loss": 5.3070969581604, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.021641355007886887, + "learning_rate": 0.0006, + "loss": 5.360584735870361, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.019899170845746994, + "learning_rate": 0.0006, + "loss": 5.398155212402344, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.02298869751393795, + "learning_rate": 0.0006, + "loss": 5.410561561584473, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.02815551869571209, + "learning_rate": 0.0006, + "loss": 5.298604965209961, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.03041219152510166, + "learning_rate": 0.0006, + "loss": 5.425869941711426, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.025710400193929672, + "learning_rate": 0.0006, + "loss": 5.344316482543945, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.018438996747136116, + "learning_rate": 0.0006, + "loss": 5.379913330078125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.016683703288435936, + "learning_rate": 0.0006, + "loss": 5.338386058807373, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.017252853140234947, + "learning_rate": 0.0006, + "loss": 5.360256195068359, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.019606001675128937, + "learning_rate": 0.0006, + "loss": 5.401415824890137, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.016299517825245857, + "learning_rate": 0.0006, + "loss": 5.39270544052124, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.016657982021570206, + "learning_rate": 0.0006, + "loss": 5.359323978424072, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.017462270334362984, + "learning_rate": 0.0006, + "loss": 5.453655242919922, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01723778434097767, + "learning_rate": 0.0006, + "loss": 5.454940319061279, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.019071755930781364, + "learning_rate": 0.0006, + "loss": 5.355914115905762, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02169782482087612, + "learning_rate": 0.0006, + "loss": 5.396806716918945, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.025049131363630295, + "learning_rate": 0.0006, + "loss": 5.347422122955322, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.024977540597319603, + "learning_rate": 0.0006, + "loss": 5.32526969909668, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.024223418906331062, + "learning_rate": 0.0006, + "loss": 5.372445106506348, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.02519913762807846, + "learning_rate": 0.0006, + "loss": 5.280451774597168, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.0255630724132061, + "learning_rate": 0.0006, + "loss": 5.462911605834961, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.022740041837096214, + "learning_rate": 0.0006, + "loss": 5.319089412689209, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02038649097084999, + "learning_rate": 0.0006, + "loss": 5.37426233291626, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.020509786903858185, + "learning_rate": 0.0006, + "loss": 5.337924957275391, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.021707357838749886, + "learning_rate": 0.0006, + "loss": 5.374853134155273, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02714136242866516, + "learning_rate": 0.0006, + "loss": 5.2982282638549805, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.027913225814700127, + "learning_rate": 0.0006, + "loss": 5.406044960021973, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.948610305786133, + "eval_runtime": 75.79, + "eval_samples_per_second": 32.221, + "eval_steps_per_second": 1.016, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.023761706426739693, + "learning_rate": 0.0006, + "loss": 5.378547191619873, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.026866815984249115, + "learning_rate": 0.0006, + "loss": 5.356947898864746, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.03335791453719139, + "learning_rate": 0.0006, + "loss": 5.264157772064209, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03153866529464722, + "learning_rate": 0.0006, + "loss": 5.363075256347656, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.03270367905497551, + "learning_rate": 0.0006, + "loss": 5.298289775848389, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.030872846022248268, + "learning_rate": 0.0006, + "loss": 5.324321269989014, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.028438350185751915, + "learning_rate": 0.0006, + "loss": 5.3512115478515625, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.030846193432807922, + "learning_rate": 0.0006, + "loss": 5.304851531982422, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.03879433870315552, + "learning_rate": 0.0006, + "loss": 5.289634704589844, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.03665967285633087, + "learning_rate": 0.0006, + "loss": 5.352825164794922, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03454619646072388, + "learning_rate": 0.0006, + "loss": 5.346499443054199, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.043245580047369, + "learning_rate": 0.0006, + "loss": 5.344263076782227, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.04450082778930664, + "learning_rate": 0.0006, + "loss": 5.447438716888428, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.04980820417404175, + "learning_rate": 0.0006, + "loss": 5.364802360534668, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.04757903143763542, + "learning_rate": 0.0006, + "loss": 5.363414287567139, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.048647016286849976, + "learning_rate": 0.0006, + "loss": 5.379604339599609, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.04497060924768448, + "learning_rate": 0.0006, + "loss": 5.354029178619385, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.032642051577568054, + "learning_rate": 0.0006, + "loss": 5.335702419281006, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.037766437977552414, + "learning_rate": 0.0006, + "loss": 5.331804275512695, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.030145324766635895, + "learning_rate": 0.0006, + "loss": 5.327176094055176, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.032248858362436295, + "learning_rate": 0.0006, + "loss": 5.2410712242126465, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026990503072738647, + "learning_rate": 0.0006, + "loss": 5.369958877563477, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.023413540795445442, + "learning_rate": 0.0006, + "loss": 5.322418212890625, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025248292833566666, + "learning_rate": 0.0006, + "loss": 5.300156116485596, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.021037699654698372, + "learning_rate": 0.0006, + "loss": 5.26978063583374, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.0155200669541955, + "learning_rate": 0.0006, + "loss": 5.297061920166016, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.017867842689156532, + "learning_rate": 0.0006, + "loss": 5.312634468078613, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.01675534062087536, + "learning_rate": 0.0006, + "loss": 5.290635108947754, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.015546303242444992, + "learning_rate": 0.0006, + "loss": 5.289699554443359, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01612856611609459, + "learning_rate": 0.0006, + "loss": 5.312010765075684, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.015225349925458431, + "learning_rate": 0.0006, + "loss": 5.255838871002197, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015317161567509174, + "learning_rate": 0.0006, + "loss": 5.234170436859131, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.014820792712271214, + "learning_rate": 0.0006, + "loss": 5.255383014678955, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01343776099383831, + "learning_rate": 0.0006, + "loss": 5.256904602050781, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.01418706402182579, + "learning_rate": 0.0006, + "loss": 5.272562026977539, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.013406263664364815, + "learning_rate": 0.0006, + "loss": 5.17042350769043, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.01331746019423008, + "learning_rate": 0.0006, + "loss": 5.267889022827148, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.014625098556280136, + "learning_rate": 0.0006, + "loss": 5.281431198120117, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01346531230956316, + "learning_rate": 0.0006, + "loss": 5.286074638366699, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01303934771567583, + "learning_rate": 0.0006, + "loss": 5.344804763793945, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.014180841855704784, + "learning_rate": 0.0006, + "loss": 5.252955436706543, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.014279021881520748, + "learning_rate": 0.0006, + "loss": 5.283501625061035, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.013922203332185745, + "learning_rate": 0.0006, + "loss": 5.225350379943848, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.014055478386580944, + "learning_rate": 0.0006, + "loss": 5.283530235290527, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.013623543083667755, + "learning_rate": 0.0006, + "loss": 5.204877853393555, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.014378432184457779, + "learning_rate": 0.0006, + "loss": 5.275735378265381, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.017884068191051483, + "learning_rate": 0.0006, + "loss": 5.256485939025879, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.0203024260699749, + "learning_rate": 0.0006, + "loss": 5.292258262634277, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.021196166053414345, + "learning_rate": 0.0006, + "loss": 5.30052375793457, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.02249213121831417, + "learning_rate": 0.0006, + "loss": 5.236924648284912, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.019980570301413536, + "learning_rate": 0.0006, + "loss": 5.290444374084473, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.022030364722013474, + "learning_rate": 0.0006, + "loss": 5.287881851196289, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.02438683621585369, + "learning_rate": 0.0006, + "loss": 5.230462074279785, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.026895932853221893, + "learning_rate": 0.0006, + "loss": 5.232667446136475, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.02861972153186798, + "learning_rate": 0.0006, + "loss": 5.205913066864014, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.033064406365156174, + "learning_rate": 0.0006, + "loss": 5.242166042327881, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.030850905925035477, + "learning_rate": 0.0006, + "loss": 5.2503581047058105, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.03139309585094452, + "learning_rate": 0.0006, + "loss": 5.222465515136719, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.035760655999183655, + "learning_rate": 0.0006, + "loss": 5.279786586761475, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.036571063101291656, + "learning_rate": 0.0006, + "loss": 5.237204551696777, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.03467971459031105, + "learning_rate": 0.0006, + "loss": 5.210154056549072, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.053154416382312775, + "learning_rate": 0.0006, + "loss": 5.295984268188477, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.05374281480908394, + "learning_rate": 0.0006, + "loss": 5.293900012969971, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.05851082503795624, + "learning_rate": 0.0006, + "loss": 5.3758931159973145, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.05145686864852905, + "learning_rate": 0.0006, + "loss": 5.309441566467285, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.04484057426452637, + "learning_rate": 0.0006, + "loss": 5.297204494476318, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.05348053574562073, + "learning_rate": 0.0006, + "loss": 5.278589248657227, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.05677935481071472, + "learning_rate": 0.0006, + "loss": 5.290632247924805, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.05244941636919975, + "learning_rate": 0.0006, + "loss": 5.353776454925537, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.03900328651070595, + "learning_rate": 0.0006, + "loss": 5.284013748168945, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.03373100981116295, + "learning_rate": 0.0006, + "loss": 5.3313751220703125, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.03497432917356491, + "learning_rate": 0.0006, + "loss": 5.395255088806152, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.799665451049805, + "eval_runtime": 81.3525, + "eval_samples_per_second": 30.018, + "eval_steps_per_second": 0.946, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.039509695023298264, + "learning_rate": 0.0006, + "loss": 5.298876762390137, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.034599535167217255, + "learning_rate": 0.0006, + "loss": 5.264098167419434, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.029622144997119904, + "learning_rate": 0.0006, + "loss": 5.21397066116333, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.02309993840754032, + "learning_rate": 0.0006, + "loss": 5.227992057800293, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.024507572874426842, + "learning_rate": 0.0006, + "loss": 5.192324161529541, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.018214358016848564, + "learning_rate": 0.0006, + "loss": 5.23871374130249, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.01937488280236721, + "learning_rate": 0.0006, + "loss": 5.227265357971191, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.017225295305252075, + "learning_rate": 0.0006, + "loss": 5.232108116149902, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.016314376145601273, + "learning_rate": 0.0006, + "loss": 5.2288360595703125, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.014701297506690025, + "learning_rate": 0.0006, + "loss": 5.297238349914551, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.015610086731612682, + "learning_rate": 0.0006, + "loss": 5.148407936096191, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.014056755229830742, + "learning_rate": 0.0006, + "loss": 5.208695411682129, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.012802829965949059, + "learning_rate": 0.0006, + "loss": 5.229208946228027, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.011958962306380272, + "learning_rate": 0.0006, + "loss": 5.201716423034668, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.012841588817536831, + "learning_rate": 0.0006, + "loss": 5.240131378173828, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.014653137885034084, + "learning_rate": 0.0006, + "loss": 5.146340370178223, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.010446185246109962, + "learning_rate": 0.0006, + "loss": 5.21809720993042, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.01092249620705843, + "learning_rate": 0.0006, + "loss": 5.135411739349365, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.010561129078269005, + "learning_rate": 0.0006, + "loss": 5.194492340087891, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.010118645615875721, + "learning_rate": 0.0006, + "loss": 5.235012531280518, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.01081101130694151, + "learning_rate": 0.0006, + "loss": 5.162468910217285, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.010701699182391167, + "learning_rate": 0.0006, + "loss": 5.217264175415039, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.010495852679014206, + "learning_rate": 0.0006, + "loss": 5.202086925506592, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.009838009253144264, + "learning_rate": 0.0006, + "loss": 5.201089859008789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.011280231177806854, + "learning_rate": 0.0006, + "loss": 5.1775336265563965, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.011244936846196651, + "learning_rate": 0.0006, + "loss": 5.185998439788818, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.01153804361820221, + "learning_rate": 0.0006, + "loss": 5.132739067077637, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010817309841513634, + "learning_rate": 0.0006, + "loss": 5.268172264099121, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.011870342306792736, + "learning_rate": 0.0006, + "loss": 5.154515266418457, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.011373576708137989, + "learning_rate": 0.0006, + "loss": 5.160059928894043, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.01116643100976944, + "learning_rate": 0.0006, + "loss": 5.079656600952148, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.01202121190726757, + "learning_rate": 0.0006, + "loss": 5.205666542053223, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.013683444820344448, + "learning_rate": 0.0006, + "loss": 5.168977737426758, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.012892362661659718, + "learning_rate": 0.0006, + "loss": 5.168679237365723, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.01123902853578329, + "learning_rate": 0.0006, + "loss": 5.143795013427734, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.010333497077226639, + "learning_rate": 0.0006, + "loss": 5.170592784881592, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.011278851889073849, + "learning_rate": 0.0006, + "loss": 5.162395477294922, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.013309221714735031, + "learning_rate": 0.0006, + "loss": 5.0712103843688965, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.014573273248970509, + "learning_rate": 0.0006, + "loss": 5.098300933837891, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.01755373552441597, + "learning_rate": 0.0006, + "loss": 5.081066131591797, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02018735185265541, + "learning_rate": 0.0006, + "loss": 5.173007965087891, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.023258069530129433, + "learning_rate": 0.0006, + "loss": 5.107330322265625, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.0250930804759264, + "learning_rate": 0.0006, + "loss": 5.193347930908203, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02333216182887554, + "learning_rate": 0.0006, + "loss": 5.160326957702637, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019765766337513924, + "learning_rate": 0.0006, + "loss": 5.05100679397583, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.020395414903759956, + "learning_rate": 0.0006, + "loss": 5.157297611236572, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019553566351532936, + "learning_rate": 0.0006, + "loss": 5.158632755279541, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.018838031217455864, + "learning_rate": 0.0006, + "loss": 5.167074203491211, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.020884843543171883, + "learning_rate": 0.0006, + "loss": 5.238006114959717, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.021462196484208107, + "learning_rate": 0.0006, + "loss": 5.1703901290893555, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.021178999915719032, + "learning_rate": 0.0006, + "loss": 5.109465599060059, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.022374693304300308, + "learning_rate": 0.0006, + "loss": 5.195014953613281, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.021914934739470482, + "learning_rate": 0.0006, + "loss": 5.15154504776001, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.017769858241081238, + "learning_rate": 0.0006, + "loss": 5.119076251983643, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.02018590457737446, + "learning_rate": 0.0006, + "loss": 5.087010860443115, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.023143045604228973, + "learning_rate": 0.0006, + "loss": 5.087374687194824, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.02214638516306877, + "learning_rate": 0.0006, + "loss": 5.043274879455566, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.024159753695130348, + "learning_rate": 0.0006, + "loss": 5.11695671081543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.023281913250684738, + "learning_rate": 0.0006, + "loss": 5.159364700317383, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.02846658229827881, + "learning_rate": 0.0006, + "loss": 5.0652570724487305, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.03165476396679878, + "learning_rate": 0.0006, + "loss": 5.136387825012207, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.02582591027021408, + "learning_rate": 0.0006, + "loss": 5.059978485107422, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.025306783616542816, + "learning_rate": 0.0006, + "loss": 5.123497486114502, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.02598070539534092, + "learning_rate": 0.0006, + "loss": 5.155566215515137, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.026244031265378, + "learning_rate": 0.0006, + "loss": 5.0871381759643555, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.025219639763236046, + "learning_rate": 0.0006, + "loss": 5.102408409118652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.023916568607091904, + "learning_rate": 0.0006, + "loss": 5.131472587585449, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.019619153812527657, + "learning_rate": 0.0006, + "loss": 5.14764404296875, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.019982660189270973, + "learning_rate": 0.0006, + "loss": 5.0941972732543945, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.018386343494057655, + "learning_rate": 0.0006, + "loss": 5.138663291931152, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.021332457661628723, + "learning_rate": 0.0006, + "loss": 5.141696453094482, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.021462492644786835, + "learning_rate": 0.0006, + "loss": 5.049698352813721, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.544618129730225, + "eval_runtime": 72.1254, + "eval_samples_per_second": 33.858, + "eval_steps_per_second": 1.068, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02191290818154812, + "learning_rate": 0.0006, + "loss": 5.0831732749938965, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02411566860973835, + "learning_rate": 0.0006, + "loss": 5.116334915161133, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.027553394436836243, + "learning_rate": 0.0006, + "loss": 5.050045013427734, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02357514388859272, + "learning_rate": 0.0006, + "loss": 5.091012001037598, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.021199429407715797, + "learning_rate": 0.0006, + "loss": 5.104509353637695, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.019254926592111588, + "learning_rate": 0.0006, + "loss": 5.141340255737305, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.01811135932803154, + "learning_rate": 0.0006, + "loss": 5.046824932098389, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.017739299684762955, + "learning_rate": 0.0006, + "loss": 5.045833587646484, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.018534071743488312, + "learning_rate": 0.0006, + "loss": 4.9618401527404785, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.016690023243427277, + "learning_rate": 0.0006, + "loss": 5.165715217590332, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.018630729988217354, + "learning_rate": 0.0006, + "loss": 5.082975387573242, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01884593442082405, + "learning_rate": 0.0006, + "loss": 5.048741817474365, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018500948324799538, + "learning_rate": 0.0006, + "loss": 5.105315685272217, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.020263809710741043, + "learning_rate": 0.0006, + "loss": 5.087735176086426, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.019454646855592728, + "learning_rate": 0.0006, + "loss": 5.0284810066223145, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.01877632364630699, + "learning_rate": 0.0006, + "loss": 5.0442609786987305, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.020227789878845215, + "learning_rate": 0.0006, + "loss": 5.081351280212402, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.018256092444062233, + "learning_rate": 0.0006, + "loss": 5.016656875610352, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01916426233947277, + "learning_rate": 0.0006, + "loss": 5.034103870391846, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.017990969121456146, + "learning_rate": 0.0006, + "loss": 5.039344310760498, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.018346179276704788, + "learning_rate": 0.0006, + "loss": 5.017010688781738, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.01584712415933609, + "learning_rate": 0.0006, + "loss": 4.992602348327637, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.014064064249396324, + "learning_rate": 0.0006, + "loss": 5.020243167877197, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.014658658765256405, + "learning_rate": 0.0006, + "loss": 4.995723724365234, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.015613174997270107, + "learning_rate": 0.0006, + "loss": 5.0587310791015625, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.016534067690372467, + "learning_rate": 0.0006, + "loss": 4.943386077880859, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.016116850078105927, + "learning_rate": 0.0006, + "loss": 5.117921829223633, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.014806007035076618, + "learning_rate": 0.0006, + "loss": 5.077454566955566, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.013742697425186634, + "learning_rate": 0.0006, + "loss": 5.059109687805176, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.01445091050118208, + "learning_rate": 0.0006, + "loss": 4.97079610824585, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.014255733229219913, + "learning_rate": 0.0006, + "loss": 5.056376934051514, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.014356110244989395, + "learning_rate": 0.0006, + "loss": 5.095856666564941, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.015831749886274338, + "learning_rate": 0.0006, + "loss": 5.063546180725098, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.01640896685421467, + "learning_rate": 0.0006, + "loss": 5.038229942321777, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.017240721732378006, + "learning_rate": 0.0006, + "loss": 4.989873886108398, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.020683523267507553, + "learning_rate": 0.0006, + "loss": 4.945237159729004, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.02303830161690712, + "learning_rate": 0.0006, + "loss": 5.039238929748535, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02372933179140091, + "learning_rate": 0.0006, + "loss": 5.005778789520264, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.02596980705857277, + "learning_rate": 0.0006, + "loss": 5.047043800354004, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.03045373223721981, + "learning_rate": 0.0006, + "loss": 5.07199764251709, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.033439598977565765, + "learning_rate": 0.0006, + "loss": 5.008430480957031, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.027381205931305885, + "learning_rate": 0.0006, + "loss": 5.10188102722168, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.029322044923901558, + "learning_rate": 0.0006, + "loss": 5.0966997146606445, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.02821085974574089, + "learning_rate": 0.0006, + "loss": 4.9426422119140625, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.03356366604566574, + "learning_rate": 0.0006, + "loss": 5.083784580230713, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.03138979524374008, + "learning_rate": 0.0006, + "loss": 5.063078880310059, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.02739642933011055, + "learning_rate": 0.0006, + "loss": 5.0485687255859375, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.026842772960662842, + "learning_rate": 0.0006, + "loss": 5.040291786193848, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.03270867466926575, + "learning_rate": 0.0006, + "loss": 5.118686199188232, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.027592353522777557, + "learning_rate": 0.0006, + "loss": 5.05806303024292, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.029291553422808647, + "learning_rate": 0.0006, + "loss": 5.048781394958496, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.027100825682282448, + "learning_rate": 0.0006, + "loss": 5.09072208404541, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.02333410270512104, + "learning_rate": 0.0006, + "loss": 5.0193939208984375, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.022513234987854958, + "learning_rate": 0.0006, + "loss": 4.96424674987793, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01899099536240101, + "learning_rate": 0.0006, + "loss": 4.982175350189209, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.0192430317401886, + "learning_rate": 0.0006, + "loss": 5.094186782836914, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02133013680577278, + "learning_rate": 0.0006, + "loss": 5.04017448425293, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.019499894231557846, + "learning_rate": 0.0006, + "loss": 5.058502197265625, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.016153644770383835, + "learning_rate": 0.0006, + "loss": 5.006505012512207, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.016342492774128914, + "learning_rate": 0.0006, + "loss": 4.998578071594238, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.014268498867750168, + "learning_rate": 0.0006, + "loss": 5.099554538726807, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.014293930493295193, + "learning_rate": 0.0006, + "loss": 4.929508686065674, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.013110974803566933, + "learning_rate": 0.0006, + "loss": 4.981067657470703, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.014049272052943707, + "learning_rate": 0.0006, + "loss": 4.97393274307251, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.011994240805506706, + "learning_rate": 0.0006, + "loss": 4.9473419189453125, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.013252398930490017, + "learning_rate": 0.0006, + "loss": 4.965734481811523, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.013359254226088524, + "learning_rate": 0.0006, + "loss": 4.968218803405762, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.01238918025046587, + "learning_rate": 0.0006, + "loss": 5.07440710067749, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.014492994174361229, + "learning_rate": 0.0006, + "loss": 5.0300493240356445, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.01550526823848486, + "learning_rate": 0.0006, + "loss": 5.031158924102783, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.016029849648475647, + "learning_rate": 0.0006, + "loss": 5.000946521759033, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.0141118373721838, + "learning_rate": 0.0006, + "loss": 5.079067230224609, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 5.463352203369141, + "eval_runtime": 79.3035, + "eval_samples_per_second": 30.793, + "eval_steps_per_second": 0.971, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.015567110851407051, + "learning_rate": 0.0006, + "loss": 5.006353855133057, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.017325397580862045, + "learning_rate": 0.0006, + "loss": 5.022473335266113, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.018784036859869957, + "learning_rate": 0.0006, + "loss": 4.922494888305664, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.020597755908966064, + "learning_rate": 0.0006, + "loss": 5.045244216918945, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.023631447926163673, + "learning_rate": 0.0006, + "loss": 4.916990280151367, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.02467048540711403, + "learning_rate": 0.0006, + "loss": 5.0631208419799805, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.024088503792881966, + "learning_rate": 0.0006, + "loss": 4.917226791381836, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.022297589108347893, + "learning_rate": 0.0006, + "loss": 4.940986156463623, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02313072420656681, + "learning_rate": 0.0006, + "loss": 5.013306617736816, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.028538871556520462, + "learning_rate": 0.0006, + "loss": 4.99459171295166, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.029122058302164078, + "learning_rate": 0.0006, + "loss": 5.000114440917969, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.022197917103767395, + "learning_rate": 0.0006, + "loss": 4.945086479187012, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.021084975451231003, + "learning_rate": 0.0006, + "loss": 4.920244216918945, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.022474950179457664, + "learning_rate": 0.0006, + "loss": 5.005110740661621, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.02403559722006321, + "learning_rate": 0.0006, + "loss": 5.04328727722168, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.026955733075737953, + "learning_rate": 0.0006, + "loss": 4.962291717529297, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.024654097855091095, + "learning_rate": 0.0006, + "loss": 5.0805840492248535, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.01992264948785305, + "learning_rate": 0.0006, + "loss": 5.044548034667969, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01696479320526123, + "learning_rate": 0.0006, + "loss": 4.912850856781006, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.017840469256043434, + "learning_rate": 0.0006, + "loss": 5.0556793212890625, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.01623854599893093, + "learning_rate": 0.0006, + "loss": 5.040976524353027, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.016552211716771126, + "learning_rate": 0.0006, + "loss": 4.971366882324219, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.015316464938223362, + "learning_rate": 0.0006, + "loss": 4.952664375305176, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01712590828537941, + "learning_rate": 0.0006, + "loss": 4.969407081604004, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.015214908868074417, + "learning_rate": 0.0006, + "loss": 4.9463911056518555, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.014866896905004978, + "learning_rate": 0.0006, + "loss": 5.071913242340088, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.01533361803740263, + "learning_rate": 0.0006, + "loss": 5.0071916580200195, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.014022483490407467, + "learning_rate": 0.0006, + "loss": 4.954067707061768, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.015019413083791733, + "learning_rate": 0.0006, + "loss": 4.9905900955200195, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.015474580228328705, + "learning_rate": 0.0006, + "loss": 4.999631881713867, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.020479638129472733, + "learning_rate": 0.0006, + "loss": 5.0200700759887695, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.0237763449549675, + "learning_rate": 0.0006, + "loss": 5.008729457855225, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.022637002170085907, + "learning_rate": 0.0006, + "loss": 4.940751075744629, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.019718417897820473, + "learning_rate": 0.0006, + "loss": 4.9788737297058105, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023200126364827156, + "learning_rate": 0.0006, + "loss": 4.906790256500244, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.025976872071623802, + "learning_rate": 0.0006, + "loss": 5.002409934997559, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.023092452436685562, + "learning_rate": 0.0006, + "loss": 4.90887451171875, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.01955891214311123, + "learning_rate": 0.0006, + "loss": 4.995199203491211, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02019096165895462, + "learning_rate": 0.0006, + "loss": 4.908695697784424, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.019376836717128754, + "learning_rate": 0.0006, + "loss": 4.872652053833008, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.019068755209445953, + "learning_rate": 0.0006, + "loss": 4.952585220336914, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.016856160014867783, + "learning_rate": 0.0006, + "loss": 4.927958965301514, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.017544621601700783, + "learning_rate": 0.0006, + "loss": 4.922074794769287, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.01814766600728035, + "learning_rate": 0.0006, + "loss": 4.884115219116211, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.01756724901497364, + "learning_rate": 0.0006, + "loss": 4.969221115112305, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.016586478799581528, + "learning_rate": 0.0006, + "loss": 4.897098541259766, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.0165987778455019, + "learning_rate": 0.0006, + "loss": 4.8751044273376465, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019033856689929962, + "learning_rate": 0.0006, + "loss": 4.894170761108398, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.016953010112047195, + "learning_rate": 0.0006, + "loss": 4.918987274169922, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016293920576572418, + "learning_rate": 0.0006, + "loss": 4.949659824371338, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.019263453781604767, + "learning_rate": 0.0006, + "loss": 4.923409461975098, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01789483241736889, + "learning_rate": 0.0006, + "loss": 4.894615173339844, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.9311604499816895, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01674770377576351, + "learning_rate": 0.0006, + "loss": 4.938290596008301, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01579914800822735, + "learning_rate": 0.0006, + "loss": 4.936288833618164, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.015871549025177956, + "learning_rate": 0.0006, + "loss": 4.970236301422119, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.014079514890909195, + "learning_rate": 0.0006, + "loss": 4.935883045196533, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.015348718501627445, + "learning_rate": 0.0006, + "loss": 4.919315814971924, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.0189143568277359, + "learning_rate": 0.0006, + "loss": 4.972513198852539, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.021286798641085625, + "learning_rate": 0.0006, + "loss": 4.89121675491333, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.0212328489869833, + "learning_rate": 0.0006, + "loss": 4.895394802093506, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.020312150940299034, + "learning_rate": 0.0006, + "loss": 4.943729400634766, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.02121325396001339, + "learning_rate": 0.0006, + "loss": 4.9322638511657715, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.02030770666897297, + "learning_rate": 0.0006, + "loss": 4.8969550132751465, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.02408582717180252, + "learning_rate": 0.0006, + "loss": 4.94023323059082, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.025230303406715393, + "learning_rate": 0.0006, + "loss": 4.964077472686768, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.025903481990098953, + "learning_rate": 0.0006, + "loss": 4.9319024085998535, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.02517629973590374, + "learning_rate": 0.0006, + "loss": 4.9937262535095215, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.023874718695878983, + "learning_rate": 0.0006, + "loss": 4.916127681732178, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.02531718835234642, + "learning_rate": 0.0006, + "loss": 4.949875831604004, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.026585623621940613, + "learning_rate": 0.0006, + "loss": 4.883786201477051, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.029191339388489723, + "learning_rate": 0.0006, + "loss": 4.955547332763672, + "step": 1152 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 4.9113813781315584e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1152/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef +size 4792 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/config.json new file mode 100644 index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151674 +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0e36a9e703f0be14f7d30556367636bd1d91e1c5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ab162fa96483f968ae7ded2cf2b482442f54e0da91166a0b597716364e7c064 +size 583362376 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..a1eb75fd2c66e5adb2d7586ade67c4886d8f6a99 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80e0b92eff1988e3f1b7e6cc77dcc94de424b81f7d783bcaa3d266463e964d8f +size 1166837626 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..b64929d57263eff2fb995a23b08eba35f6f93010 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63caf9dac92083e4136aebe27ed5f14192e67e2b7c1be28ddb92c9378e30cbec +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..4c39af5c7e36ba4e1c7db044f83e67f098dcb23a --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:865832ee592ced5b17b41bdfc0ab0aa506de831285ad28234b98fa4dcfae8116 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..607d15463fe50c2ab955ee306939c8c8e89b2d37 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cc13de805c6c2ec23231d41507998e68cb38eb124f7e877be15ee1e574b7c57 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..dab9765a6a5ab1ac06d19e29f8f7d3e1c80aaedd --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:654b8fd45d1880a532d0df9f8d24dd859c9a20985b2bc449c6d9ebcffb476338 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..0d34fe10b069b50b5fecbd4a0affbd8f653b7b9b --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a84bbc84bb0dd489a2d72fd441cc0b8482e9e06d7cde3ac44b50a389883ce86 +size 1064 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14 +size 11424648 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/tokenizer_config.json @@ -0,0 +1,21 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>", + "<|mask|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..61823ff1b5d55f798d5b62cbe1d5f3533f3ec8a1 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/trainer_state.json @@ -0,0 +1,10266 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 20.0, + "eval_steps": 500, + "global_step": 1440, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.5768774151802063, + "learning_rate": 0.0, + "loss": 12.050328254699707, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.5784570574760437, + "learning_rate": 5.999999999999999e-06, + "loss": 12.049251556396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.5607187151908875, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.921792030334473, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.46466919779777527, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.686628341674805, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.38012945652008057, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.432823181152344, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.33215776085853577, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.207542419433594, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.2928776741027832, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.000097274780273, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.26018962264060974, + "learning_rate": 4.2e-05, + "loss": 10.825748443603516, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.23717083036899567, + "learning_rate": 4.7999999999999994e-05, + "loss": 10.677122116088867, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.21641355752944946, + "learning_rate": 5.399999999999999e-05, + "loss": 10.569148063659668, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.1982956826686859, + "learning_rate": 5.9999999999999995e-05, + "loss": 10.49653434753418, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.18819282948970795, + "learning_rate": 6.599999999999999e-05, + "loss": 10.432951927185059, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.18322618305683136, + "learning_rate": 7.199999999999999e-05, + "loss": 10.387115478515625, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.18399226665496826, + "learning_rate": 7.8e-05, + "loss": 10.33381462097168, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.1833190619945526, + "learning_rate": 8.4e-05, + "loss": 10.301689147949219, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.18490834534168243, + "learning_rate": 8.999999999999999e-05, + "loss": 10.240968704223633, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.18358023464679718, + "learning_rate": 9.599999999999999e-05, + "loss": 10.194717407226562, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.1801682710647583, + "learning_rate": 0.000102, + "loss": 10.14897632598877, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.17899440228939056, + "learning_rate": 0.00010799999999999998, + "loss": 10.078763961791992, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.17666497826576233, + "learning_rate": 0.00011399999999999999, + "loss": 10.01095962524414, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.17196154594421387, + "learning_rate": 0.00011999999999999999, + "loss": 9.951203346252441, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.169203519821167, + "learning_rate": 0.00012599999999999997, + "loss": 9.871726036071777, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.16500306129455566, + "learning_rate": 0.00013199999999999998, + "loss": 9.803646087646484, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.16128124296665192, + "learning_rate": 0.000138, + "loss": 9.728487014770508, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.15904058516025543, + "learning_rate": 0.00014399999999999998, + "loss": 9.642473220825195, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.15560561418533325, + "learning_rate": 0.00015, + "loss": 9.56414794921875, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.15278612077236176, + "learning_rate": 0.000156, + "loss": 9.484230041503906, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.15263237059116364, + "learning_rate": 0.000162, + "loss": 9.383644104003906, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.14945024251937866, + "learning_rate": 0.000168, + "loss": 9.29840087890625, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.14614646136760712, + "learning_rate": 0.00017399999999999997, + "loss": 9.208931922912598, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.14295098185539246, + "learning_rate": 0.00017999999999999998, + "loss": 9.122323989868164, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.13868345320224762, + "learning_rate": 0.000186, + "loss": 9.03034496307373, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.135064035654068, + "learning_rate": 0.00019199999999999998, + "loss": 8.92818832397461, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.13097241520881653, + "learning_rate": 0.000198, + "loss": 8.834627151489258, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.12478776276111603, + "learning_rate": 0.000204, + "loss": 8.755172729492188, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.1214982345700264, + "learning_rate": 0.00020999999999999998, + "loss": 8.647891998291016, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.11958328634500504, + "learning_rate": 0.00021599999999999996, + "loss": 8.543274879455566, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.11680915951728821, + "learning_rate": 0.00022199999999999998, + "loss": 8.44540023803711, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.11491474509239197, + "learning_rate": 0.00022799999999999999, + "loss": 8.349921226501465, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.11160755902528763, + "learning_rate": 0.000234, + "loss": 8.26960563659668, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.10840637236833572, + "learning_rate": 0.00023999999999999998, + "loss": 8.192136764526367, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.10585298389196396, + "learning_rate": 0.00024599999999999996, + "loss": 8.09546947479248, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.1032542735338211, + "learning_rate": 0.00025199999999999995, + "loss": 7.991090774536133, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.0986943170428276, + "learning_rate": 0.000258, + "loss": 7.92384147644043, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.09320762753486633, + "learning_rate": 0.00026399999999999997, + "loss": 7.8478193283081055, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.09168332070112228, + "learning_rate": 0.00027, + "loss": 7.76185417175293, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.08832395076751709, + "learning_rate": 0.000276, + "loss": 7.679399490356445, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.08816400915384293, + "learning_rate": 0.00028199999999999997, + "loss": 7.5789642333984375, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0861717164516449, + "learning_rate": 0.00028799999999999995, + "loss": 7.519214630126953, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.08103746175765991, + "learning_rate": 0.000294, + "loss": 7.4475321769714355, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.07545104622840881, + "learning_rate": 0.0003, + "loss": 7.401182174682617, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.07042575627565384, + "learning_rate": 0.00030599999999999996, + "loss": 7.340173244476318, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.06737572699785233, + "learning_rate": 0.000312, + "loss": 7.265697002410889, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.06151146441698074, + "learning_rate": 0.000318, + "loss": 7.231061935424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.05638071522116661, + "learning_rate": 0.000324, + "loss": 7.195466995239258, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.05084463208913803, + "learning_rate": 0.00033, + "loss": 7.137603759765625, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.04293261095881462, + "learning_rate": 0.000336, + "loss": 7.115458965301514, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.036064863204956055, + "learning_rate": 0.00034199999999999996, + "loss": 7.089038848876953, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.029026754200458527, + "learning_rate": 0.00034799999999999995, + "loss": 7.0584306716918945, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.02286517433822155, + "learning_rate": 0.00035399999999999993, + "loss": 7.049450874328613, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.018878433853387833, + "learning_rate": 0.00035999999999999997, + "loss": 7.008824348449707, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.010773813351988792, + "learning_rate": 0.00036599999999999995, + "loss": 7.024829864501953, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.008794732391834259, + "learning_rate": 0.000372, + "loss": 7.018399238586426, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.011709796264767647, + "learning_rate": 0.00037799999999999997, + "loss": 7.0101752281188965, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.012123258784413338, + "learning_rate": 0.00038399999999999996, + "loss": 7.024317741394043, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01530320756137371, + "learning_rate": 0.00039, + "loss": 7.027738571166992, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.01890002004802227, + "learning_rate": 0.000396, + "loss": 7.024099349975586, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.020684150978922844, + "learning_rate": 0.000402, + "loss": 7.007943153381348, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.019826196134090424, + "learning_rate": 0.000408, + "loss": 7.006874084472656, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.019218673929572105, + "learning_rate": 0.0004139999999999999, + "loss": 7.002897262573242, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.013856411911547184, + "learning_rate": 0.00041999999999999996, + "loss": 6.993999481201172, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.010390003211796284, + "learning_rate": 0.00042599999999999995, + "loss": 7.000683784484863, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.91140079498291, + "eval_runtime": 57.7925, + "eval_samples_per_second": 42.255, + "eval_steps_per_second": 1.332, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.013802163302898407, + "learning_rate": 0.00043199999999999993, + "loss": 6.998257637023926, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.010109687224030495, + "learning_rate": 0.00043799999999999997, + "loss": 7.002542018890381, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.008019139990210533, + "learning_rate": 0.00044399999999999995, + "loss": 6.985941410064697, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.0069672465324401855, + "learning_rate": 0.00045, + "loss": 6.99245023727417, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.00714076729491353, + "learning_rate": 0.00045599999999999997, + "loss": 6.959296226501465, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.008408435620367527, + "learning_rate": 0.00046199999999999995, + "loss": 6.954504489898682, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.008778786286711693, + "learning_rate": 0.000468, + "loss": 7.0007004737854, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.007753407116979361, + "learning_rate": 0.000474, + "loss": 6.98183536529541, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.01087260153144598, + "learning_rate": 0.00047999999999999996, + "loss": 6.991066932678223, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.006122548598796129, + "learning_rate": 0.000486, + "loss": 6.978824138641357, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.0067383465357124805, + "learning_rate": 0.0004919999999999999, + "loss": 6.961703777313232, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.007922791875898838, + "learning_rate": 0.000498, + "loss": 6.954860687255859, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.00818922370672226, + "learning_rate": 0.0005039999999999999, + "loss": 6.945961952209473, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.006791461259126663, + "learning_rate": 0.0005099999999999999, + "loss": 6.968680381774902, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.005394163075834513, + "learning_rate": 0.000516, + "loss": 6.956404685974121, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.004500220064073801, + "learning_rate": 0.000522, + "loss": 6.96998405456543, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008699574507772923, + "learning_rate": 0.0005279999999999999, + "loss": 6.955585479736328, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.005889365915209055, + "learning_rate": 0.000534, + "loss": 6.9708051681518555, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009646810591220856, + "learning_rate": 0.00054, + "loss": 6.962608814239502, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.004517171997576952, + "learning_rate": 0.0005459999999999999, + "loss": 6.957627296447754, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.006339677143841982, + "learning_rate": 0.000552, + "loss": 6.963141441345215, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.005387383047491312, + "learning_rate": 0.000558, + "loss": 6.934465408325195, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.005252422299236059, + "learning_rate": 0.0005639999999999999, + "loss": 6.956108093261719, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.008699414320290089, + "learning_rate": 0.00057, + "loss": 6.973369598388672, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.003947409335523844, + "learning_rate": 0.0005759999999999999, + "loss": 6.950225830078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.00768681475892663, + "learning_rate": 0.0005819999999999999, + "loss": 6.964739799499512, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.006563829258084297, + "learning_rate": 0.000588, + "loss": 6.972421169281006, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007606555242091417, + "learning_rate": 0.0005939999999999999, + "loss": 6.958213806152344, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.006250880658626556, + "learning_rate": 0.0006, + "loss": 6.962296485900879, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.0073033408261835575, + "learning_rate": 0.0006, + "loss": 6.969705104827881, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.0054510910995304585, + "learning_rate": 0.0006, + "loss": 6.9652862548828125, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.004895673133432865, + "learning_rate": 0.0006, + "loss": 6.956023693084717, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.005306515377014875, + "learning_rate": 0.0006, + "loss": 6.963231086730957, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.004742146469652653, + "learning_rate": 0.0006, + "loss": 6.9582061767578125, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.004220000468194485, + "learning_rate": 0.0006, + "loss": 6.972961902618408, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0042153168469667435, + "learning_rate": 0.0006, + "loss": 6.961165904998779, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.005165040958672762, + "learning_rate": 0.0006, + "loss": 6.95018196105957, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.003930266480892897, + "learning_rate": 0.0006, + "loss": 6.9651689529418945, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.004465071018785238, + "learning_rate": 0.0006, + "loss": 6.960553169250488, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.004888612311333418, + "learning_rate": 0.0006, + "loss": 6.967258930206299, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.007497863378375769, + "learning_rate": 0.0006, + "loss": 6.954545021057129, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.005659648682922125, + "learning_rate": 0.0006, + "loss": 6.962299823760986, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.006641396787017584, + "learning_rate": 0.0006, + "loss": 6.952983856201172, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.009715928696095943, + "learning_rate": 0.0006, + "loss": 6.958622932434082, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.012789523229002953, + "learning_rate": 0.0006, + "loss": 6.932043075561523, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021673237904906273, + "learning_rate": 0.0006, + "loss": 6.931290149688721, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.16024991869926453, + "learning_rate": 0.0006, + "loss": 6.979497909545898, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.2434523105621338, + "learning_rate": 0.0006, + "loss": 7.185218811035156, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.39252421259880066, + "learning_rate": 0.0006, + "loss": 7.137226581573486, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.10565312206745148, + "learning_rate": 0.0006, + "loss": 6.984494209289551, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.1033431813120842, + "learning_rate": 0.0006, + "loss": 7.001287937164307, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.14875490963459015, + "learning_rate": 0.0006, + "loss": 7.051436901092529, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.1374562531709671, + "learning_rate": 0.0006, + "loss": 7.0293121337890625, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.0910211056470871, + "learning_rate": 0.0006, + "loss": 7.007458686828613, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.021633487194776535, + "learning_rate": 0.0006, + "loss": 6.996816635131836, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.08692754060029984, + "learning_rate": 0.0006, + "loss": 6.998541831970215, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.12613292038440704, + "learning_rate": 0.0006, + "loss": 6.9970703125, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.09462810307741165, + "learning_rate": 0.0006, + "loss": 6.997045040130615, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.03534962609410286, + "learning_rate": 0.0006, + "loss": 6.9738450050354, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.033839885145425797, + "learning_rate": 0.0006, + "loss": 6.946001052856445, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.06263269484043121, + "learning_rate": 0.0006, + "loss": 6.968981742858887, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07117997109889984, + "learning_rate": 0.0006, + "loss": 6.981535911560059, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.06336277723312378, + "learning_rate": 0.0006, + "loss": 6.976946830749512, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047467734664678574, + "learning_rate": 0.0006, + "loss": 6.961673736572266, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.018787868320941925, + "learning_rate": 0.0006, + "loss": 6.948919296264648, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.025909408926963806, + "learning_rate": 0.0006, + "loss": 6.976190567016602, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.052842333912849426, + "learning_rate": 0.0006, + "loss": 6.971207618713379, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.06421888619661331, + "learning_rate": 0.0006, + "loss": 6.979109764099121, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.055596910417079926, + "learning_rate": 0.0006, + "loss": 6.972118854522705, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.03267091140151024, + "learning_rate": 0.0006, + "loss": 6.947260856628418, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.008032613433897495, + "learning_rate": 0.0006, + "loss": 6.936816215515137, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.020184025168418884, + "learning_rate": 0.0006, + "loss": 6.929848670959473, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.836523056030273, + "eval_runtime": 58.5622, + "eval_samples_per_second": 41.699, + "eval_steps_per_second": 1.315, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.031980376690626144, + "learning_rate": 0.0006, + "loss": 6.968081474304199, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03802533075213432, + "learning_rate": 0.0006, + "loss": 6.95728874206543, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03796226158738136, + "learning_rate": 0.0006, + "loss": 6.96516752243042, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.02621951326727867, + "learning_rate": 0.0006, + "loss": 6.956058502197266, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.013292718678712845, + "learning_rate": 0.0006, + "loss": 6.9783935546875, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.009446968324482441, + "learning_rate": 0.0006, + "loss": 6.96873664855957, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.021703092381358147, + "learning_rate": 0.0006, + "loss": 6.959234237670898, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.027765290811657906, + "learning_rate": 0.0006, + "loss": 6.968524932861328, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.028761165216565132, + "learning_rate": 0.0006, + "loss": 6.954620838165283, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.019729383289813995, + "learning_rate": 0.0006, + "loss": 6.949009895324707, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.008990089409053326, + "learning_rate": 0.0006, + "loss": 6.956226348876953, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.00549644511193037, + "learning_rate": 0.0006, + "loss": 6.957143783569336, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.014436670579016209, + "learning_rate": 0.0006, + "loss": 6.954257011413574, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.01990494504570961, + "learning_rate": 0.0006, + "loss": 6.942131519317627, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.019155776128172874, + "learning_rate": 0.0006, + "loss": 6.946969985961914, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.014970783144235611, + "learning_rate": 0.0006, + "loss": 6.957938194274902, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.009875659830868244, + "learning_rate": 0.0006, + "loss": 6.953309059143066, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.004792087245732546, + "learning_rate": 0.0006, + "loss": 6.975146770477295, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.009936527349054813, + "learning_rate": 0.0006, + "loss": 6.961674690246582, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.015100127086043358, + "learning_rate": 0.0006, + "loss": 6.968759536743164, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01449542585760355, + "learning_rate": 0.0006, + "loss": 6.966510772705078, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.01073537115007639, + "learning_rate": 0.0006, + "loss": 6.936373710632324, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.006766272243112326, + "learning_rate": 0.0006, + "loss": 6.96397590637207, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.005055439658463001, + "learning_rate": 0.0006, + "loss": 6.961050987243652, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.006252622697502375, + "learning_rate": 0.0006, + "loss": 6.980506420135498, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.008084346540272236, + "learning_rate": 0.0006, + "loss": 6.9346604347229, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.010260380804538727, + "learning_rate": 0.0006, + "loss": 6.964118957519531, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.010697875171899796, + "learning_rate": 0.0006, + "loss": 6.941572189331055, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.007338830269873142, + "learning_rate": 0.0006, + "loss": 6.950471878051758, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.005931832827627659, + "learning_rate": 0.0006, + "loss": 6.962935924530029, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.0061243255622684956, + "learning_rate": 0.0006, + "loss": 6.946418285369873, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.007983328774571419, + "learning_rate": 0.0006, + "loss": 6.959758758544922, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.010470672510564327, + "learning_rate": 0.0006, + "loss": 6.93851375579834, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.011147388257086277, + "learning_rate": 0.0006, + "loss": 6.93258810043335, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01058544497936964, + "learning_rate": 0.0006, + "loss": 6.932259559631348, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.016850745305418968, + "learning_rate": 0.0006, + "loss": 6.924542427062988, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.023101547732949257, + "learning_rate": 0.0006, + "loss": 6.914878845214844, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.015795735642313957, + "learning_rate": 0.0006, + "loss": 6.922432899475098, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.008821426890790462, + "learning_rate": 0.0006, + "loss": 6.924562454223633, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.012502548284828663, + "learning_rate": 0.0006, + "loss": 6.914718151092529, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.011545066721737385, + "learning_rate": 0.0006, + "loss": 6.915705680847168, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.022190893068909645, + "learning_rate": 0.0006, + "loss": 6.88851261138916, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.023395851254463196, + "learning_rate": 0.0006, + "loss": 6.877655029296875, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015601156279444695, + "learning_rate": 0.0006, + "loss": 6.889644145965576, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01362934336066246, + "learning_rate": 0.0006, + "loss": 6.891547203063965, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.012305490672588348, + "learning_rate": 0.0006, + "loss": 6.890357971191406, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.01623358018696308, + "learning_rate": 0.0006, + "loss": 6.849443435668945, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.018063291907310486, + "learning_rate": 0.0006, + "loss": 6.842372894287109, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.010179434902966022, + "learning_rate": 0.0006, + "loss": 6.870055198669434, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.009986722841858864, + "learning_rate": 0.0006, + "loss": 6.859061241149902, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.01117654051631689, + "learning_rate": 0.0006, + "loss": 6.861199378967285, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.008551654405891895, + "learning_rate": 0.0006, + "loss": 6.850255012512207, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.018904326483607292, + "learning_rate": 0.0006, + "loss": 6.848639965057373, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.02080724574625492, + "learning_rate": 0.0006, + "loss": 6.80456018447876, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.012010873295366764, + "learning_rate": 0.0006, + "loss": 6.848775863647461, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.008113056421279907, + "learning_rate": 0.0006, + "loss": 6.849387168884277, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.012912592850625515, + "learning_rate": 0.0006, + "loss": 6.832240104675293, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.01146865077316761, + "learning_rate": 0.0006, + "loss": 6.850383281707764, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.0171892661601305, + "learning_rate": 0.0006, + "loss": 6.839577674865723, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.012083170004189014, + "learning_rate": 0.0006, + "loss": 6.818614482879639, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.0062555656768381596, + "learning_rate": 0.0006, + "loss": 6.831291198730469, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.010102466680109501, + "learning_rate": 0.0006, + "loss": 6.846522808074951, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.009042087942361832, + "learning_rate": 0.0006, + "loss": 6.822072982788086, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.011133556254208088, + "learning_rate": 0.0006, + "loss": 6.811252593994141, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.00990577507764101, + "learning_rate": 0.0006, + "loss": 6.814435958862305, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.006605073343962431, + "learning_rate": 0.0006, + "loss": 6.831300735473633, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.007831175811588764, + "learning_rate": 0.0006, + "loss": 6.841738700866699, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.005536255426704884, + "learning_rate": 0.0006, + "loss": 6.8043670654296875, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.006532667670398951, + "learning_rate": 0.0006, + "loss": 6.8091840744018555, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.006674565840512514, + "learning_rate": 0.0006, + "loss": 6.811327934265137, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.01150344219058752, + "learning_rate": 0.0006, + "loss": 6.810762405395508, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.009015528485178947, + "learning_rate": 0.0006, + "loss": 6.836431980133057, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.718832969665527, + "eval_runtime": 58.6615, + "eval_samples_per_second": 41.629, + "eval_steps_per_second": 1.313, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.009791829623281956, + "learning_rate": 0.0006, + "loss": 6.8102593421936035, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.008970947004854679, + "learning_rate": 0.0006, + "loss": 6.792663097381592, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.013689392246305943, + "learning_rate": 0.0006, + "loss": 6.815651893615723, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.031499531120061874, + "learning_rate": 0.0006, + "loss": 6.809841156005859, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03203094005584717, + "learning_rate": 0.0006, + "loss": 6.779417037963867, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.007947996258735657, + "learning_rate": 0.0006, + "loss": 6.773656845092773, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03152519091963768, + "learning_rate": 0.0006, + "loss": 6.804083824157715, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.014937258325517178, + "learning_rate": 0.0006, + "loss": 6.792213439941406, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.020961705595254898, + "learning_rate": 0.0006, + "loss": 6.8092803955078125, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.009152988903224468, + "learning_rate": 0.0006, + "loss": 6.7737603187561035, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.02770620957016945, + "learning_rate": 0.0006, + "loss": 6.767014503479004, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.01165871974080801, + "learning_rate": 0.0006, + "loss": 6.760884761810303, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.01910376362502575, + "learning_rate": 0.0006, + "loss": 6.775722980499268, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.010413747280836105, + "learning_rate": 0.0006, + "loss": 6.779816627502441, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.01918417401611805, + "learning_rate": 0.0006, + "loss": 6.760381698608398, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.017960334196686745, + "learning_rate": 0.0006, + "loss": 6.761198997497559, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.016546795144677162, + "learning_rate": 0.0006, + "loss": 6.742448806762695, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.01889597810804844, + "learning_rate": 0.0006, + "loss": 6.758237838745117, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.009611992165446281, + "learning_rate": 0.0006, + "loss": 6.75614070892334, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02034934051334858, + "learning_rate": 0.0006, + "loss": 6.736965179443359, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.010740065947175026, + "learning_rate": 0.0006, + "loss": 6.746947288513184, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.018383340910077095, + "learning_rate": 0.0006, + "loss": 6.747332572937012, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.012362911365926266, + "learning_rate": 0.0006, + "loss": 6.767669200897217, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.01127463299781084, + "learning_rate": 0.0006, + "loss": 6.769808292388916, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.017693769186735153, + "learning_rate": 0.0006, + "loss": 6.733658790588379, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.006420376244932413, + "learning_rate": 0.0006, + "loss": 6.74934720993042, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.014020364731550217, + "learning_rate": 0.0006, + "loss": 6.728157043457031, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.008018582127988338, + "learning_rate": 0.0006, + "loss": 6.730726718902588, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.00782222580164671, + "learning_rate": 0.0006, + "loss": 6.7499895095825195, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.008278409950435162, + "learning_rate": 0.0006, + "loss": 6.712803363800049, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.009826759807765484, + "learning_rate": 0.0006, + "loss": 6.742262840270996, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.009488224983215332, + "learning_rate": 0.0006, + "loss": 6.739130020141602, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.007667903788387775, + "learning_rate": 0.0006, + "loss": 6.747515678405762, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.0073090000078082085, + "learning_rate": 0.0006, + "loss": 6.740870475769043, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.010058708488941193, + "learning_rate": 0.0006, + "loss": 6.777563095092773, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.020676063373684883, + "learning_rate": 0.0006, + "loss": 6.708603858947754, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.027210544794797897, + "learning_rate": 0.0006, + "loss": 6.694097995758057, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.038053520023822784, + "learning_rate": 0.0006, + "loss": 6.697017669677734, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04684312269091606, + "learning_rate": 0.0006, + "loss": 6.729883193969727, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03673060983419418, + "learning_rate": 0.0006, + "loss": 6.743753433227539, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.006882632151246071, + "learning_rate": 0.0006, + "loss": 6.71405029296875, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.03588201105594635, + "learning_rate": 0.0006, + "loss": 6.712467193603516, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.029415199533104897, + "learning_rate": 0.0006, + "loss": 6.715634822845459, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.011425039730966091, + "learning_rate": 0.0006, + "loss": 6.697334289550781, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.0297263003885746, + "learning_rate": 0.0006, + "loss": 6.724848747253418, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.007282314356416464, + "learning_rate": 0.0006, + "loss": 6.702306270599365, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.02673051692545414, + "learning_rate": 0.0006, + "loss": 6.72357177734375, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.016174081712961197, + "learning_rate": 0.0006, + "loss": 6.730698585510254, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.0182416420429945, + "learning_rate": 0.0006, + "loss": 6.717386245727539, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.023250866681337357, + "learning_rate": 0.0006, + "loss": 6.697832107543945, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.007539753336459398, + "learning_rate": 0.0006, + "loss": 6.714348793029785, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.030086630955338478, + "learning_rate": 0.0006, + "loss": 6.724481582641602, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.012784913182258606, + "learning_rate": 0.0006, + "loss": 6.703491687774658, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.02042769268155098, + "learning_rate": 0.0006, + "loss": 6.721139907836914, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.023557329550385475, + "learning_rate": 0.0006, + "loss": 6.728122711181641, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.007716621737927198, + "learning_rate": 0.0006, + "loss": 6.70085334777832, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022772278636693954, + "learning_rate": 0.0006, + "loss": 6.712882995605469, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.007928933948278427, + "learning_rate": 0.0006, + "loss": 6.698320388793945, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.022355245426297188, + "learning_rate": 0.0006, + "loss": 6.687088489532471, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.02037850394845009, + "learning_rate": 0.0006, + "loss": 6.676856994628906, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.012290362268686295, + "learning_rate": 0.0006, + "loss": 6.682883262634277, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.022331275045871735, + "learning_rate": 0.0006, + "loss": 6.694752216339111, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.007527181413024664, + "learning_rate": 0.0006, + "loss": 6.686394691467285, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.02159409038722515, + "learning_rate": 0.0006, + "loss": 6.697943687438965, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.015054763294756413, + "learning_rate": 0.0006, + "loss": 6.677796363830566, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01466019731014967, + "learning_rate": 0.0006, + "loss": 6.697715759277344, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.01875465176999569, + "learning_rate": 0.0006, + "loss": 6.700228691101074, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.00778991961851716, + "learning_rate": 0.0006, + "loss": 6.69331693649292, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.020182441920042038, + "learning_rate": 0.0006, + "loss": 6.672651290893555, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.01240516733378172, + "learning_rate": 0.0006, + "loss": 6.688860893249512, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.011218412779271603, + "learning_rate": 0.0006, + "loss": 6.669534683227539, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.0141852255910635, + "learning_rate": 0.0006, + "loss": 6.690206527709961, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 7.537620544433594, + "eval_runtime": 58.4398, + "eval_samples_per_second": 41.787, + "eval_steps_per_second": 1.318, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.008106146939098835, + "learning_rate": 0.0006, + "loss": 6.688591003417969, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.014588244259357452, + "learning_rate": 0.0006, + "loss": 6.684961318969727, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.011985368095338345, + "learning_rate": 0.0006, + "loss": 6.672794342041016, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.014632102102041245, + "learning_rate": 0.0006, + "loss": 6.702736854553223, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.011243713088333607, + "learning_rate": 0.0006, + "loss": 6.658792495727539, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.010519202798604965, + "learning_rate": 0.0006, + "loss": 6.662495136260986, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.007318088319152594, + "learning_rate": 0.0006, + "loss": 6.675954818725586, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.011712398380041122, + "learning_rate": 0.0006, + "loss": 6.673973083496094, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.006726635154336691, + "learning_rate": 0.0006, + "loss": 6.656833171844482, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.005466983187943697, + "learning_rate": 0.0006, + "loss": 6.675468921661377, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.007163621485233307, + "learning_rate": 0.0006, + "loss": 6.6862874031066895, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.007626629434525967, + "learning_rate": 0.0006, + "loss": 6.654741287231445, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.0066330647096037865, + "learning_rate": 0.0006, + "loss": 6.68114709854126, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.007363482844084501, + "learning_rate": 0.0006, + "loss": 6.683980941772461, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.00971233006566763, + "learning_rate": 0.0006, + "loss": 6.669700622558594, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.013185256160795689, + "learning_rate": 0.0006, + "loss": 6.686490535736084, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.019116027280688286, + "learning_rate": 0.0006, + "loss": 6.645479202270508, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0328814722597599, + "learning_rate": 0.0006, + "loss": 6.68080997467041, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04482287913560867, + "learning_rate": 0.0006, + "loss": 6.698016166687012, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.03053313121199608, + "learning_rate": 0.0006, + "loss": 6.642240524291992, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.043351784348487854, + "learning_rate": 0.0006, + "loss": 6.662343978881836, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.03165886923670769, + "learning_rate": 0.0006, + "loss": 6.6291961669921875, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.028601787984371185, + "learning_rate": 0.0006, + "loss": 6.667544364929199, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.019558480009436607, + "learning_rate": 0.0006, + "loss": 6.641025543212891, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.039851363748311996, + "learning_rate": 0.0006, + "loss": 6.665952205657959, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.018417146056890488, + "learning_rate": 0.0006, + "loss": 6.657257556915283, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.02894527092576027, + "learning_rate": 0.0006, + "loss": 6.656288146972656, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.01910966821014881, + "learning_rate": 0.0006, + "loss": 6.688958168029785, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.01668412610888481, + "learning_rate": 0.0006, + "loss": 6.64876127243042, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02935275249183178, + "learning_rate": 0.0006, + "loss": 6.626096725463867, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.017997972667217255, + "learning_rate": 0.0006, + "loss": 6.640386581420898, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.01779189519584179, + "learning_rate": 0.0006, + "loss": 6.641165256500244, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.013901712372899055, + "learning_rate": 0.0006, + "loss": 6.651627063751221, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013211199082434177, + "learning_rate": 0.0006, + "loss": 6.630064964294434, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.02324107475578785, + "learning_rate": 0.0006, + "loss": 6.670443058013916, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.019452504813671112, + "learning_rate": 0.0006, + "loss": 6.651266098022461, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.017943572252988815, + "learning_rate": 0.0006, + "loss": 6.657943248748779, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.012136237695813179, + "learning_rate": 0.0006, + "loss": 6.640082359313965, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.011544011533260345, + "learning_rate": 0.0006, + "loss": 6.6186933517456055, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01569426618516445, + "learning_rate": 0.0006, + "loss": 6.640767574310303, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.027668394148349762, + "learning_rate": 0.0006, + "loss": 6.651412487030029, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.04073230177164078, + "learning_rate": 0.0006, + "loss": 6.636565208435059, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.04942672327160835, + "learning_rate": 0.0006, + "loss": 6.667994499206543, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.037568598985672, + "learning_rate": 0.0006, + "loss": 6.632008075714111, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.009880400262773037, + "learning_rate": 0.0006, + "loss": 6.6504058837890625, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.033384259790182114, + "learning_rate": 0.0006, + "loss": 6.63852596282959, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.035640884190797806, + "learning_rate": 0.0006, + "loss": 6.627986431121826, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.013637864030897617, + "learning_rate": 0.0006, + "loss": 6.624503135681152, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.025304675102233887, + "learning_rate": 0.0006, + "loss": 6.612077236175537, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.03377160802483559, + "learning_rate": 0.0006, + "loss": 6.608938217163086, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.019803153350949287, + "learning_rate": 0.0006, + "loss": 6.636343955993652, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.018928129225969315, + "learning_rate": 0.0006, + "loss": 6.636613845825195, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.023068996146321297, + "learning_rate": 0.0006, + "loss": 6.6258440017700195, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.016849957406520844, + "learning_rate": 0.0006, + "loss": 6.638483047485352, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.021430544555187225, + "learning_rate": 0.0006, + "loss": 6.619649410247803, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.024053368717432022, + "learning_rate": 0.0006, + "loss": 6.639214515686035, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.011926693841814995, + "learning_rate": 0.0006, + "loss": 6.604836463928223, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01961611397564411, + "learning_rate": 0.0006, + "loss": 6.624864101409912, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.027946067973971367, + "learning_rate": 0.0006, + "loss": 6.621284484863281, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.02108592912554741, + "learning_rate": 0.0006, + "loss": 6.5982465744018555, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.017272260040044785, + "learning_rate": 0.0006, + "loss": 6.626142501831055, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.015654582530260086, + "learning_rate": 0.0006, + "loss": 6.621735572814941, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.015797540545463562, + "learning_rate": 0.0006, + "loss": 6.6016411781311035, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.028488904237747192, + "learning_rate": 0.0006, + "loss": 6.598413467407227, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.03411734476685524, + "learning_rate": 0.0006, + "loss": 6.597321033477783, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.027188263833522797, + "learning_rate": 0.0006, + "loss": 6.630616188049316, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.03411437198519707, + "learning_rate": 0.0006, + "loss": 6.57647705078125, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.019907327368855476, + "learning_rate": 0.0006, + "loss": 6.579545974731445, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.02485930174589157, + "learning_rate": 0.0006, + "loss": 6.591073036193848, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.026106052100658417, + "learning_rate": 0.0006, + "loss": 6.580729961395264, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.029401391744613647, + "learning_rate": 0.0006, + "loss": 6.5460076332092285, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.032869379967451096, + "learning_rate": 0.0006, + "loss": 6.583278179168701, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 7.416288375854492, + "eval_runtime": 60.4177, + "eval_samples_per_second": 40.419, + "eval_steps_per_second": 1.274, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.04862433671951294, + "learning_rate": 0.0006, + "loss": 6.58473014831543, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.06636223942041397, + "learning_rate": 0.0006, + "loss": 6.621153354644775, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.05888264626264572, + "learning_rate": 0.0006, + "loss": 6.621196269989014, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.05331697314977646, + "learning_rate": 0.0006, + "loss": 6.600763320922852, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03947671875357628, + "learning_rate": 0.0006, + "loss": 6.600671768188477, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.04045343026518822, + "learning_rate": 0.0006, + "loss": 6.63529109954834, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.029766036197543144, + "learning_rate": 0.0006, + "loss": 6.612868309020996, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.03416327014565468, + "learning_rate": 0.0006, + "loss": 6.582240581512451, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03458515554666519, + "learning_rate": 0.0006, + "loss": 6.604381561279297, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.024099089205265045, + "learning_rate": 0.0006, + "loss": 6.597355365753174, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.02694438211619854, + "learning_rate": 0.0006, + "loss": 6.597512245178223, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024201856926083565, + "learning_rate": 0.0006, + "loss": 6.564685821533203, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.015959380194544792, + "learning_rate": 0.0006, + "loss": 6.582263946533203, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.025690171867609024, + "learning_rate": 0.0006, + "loss": 6.601057052612305, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.014497052878141403, + "learning_rate": 0.0006, + "loss": 6.584796905517578, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.019870297983288765, + "learning_rate": 0.0006, + "loss": 6.574010372161865, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.01807512901723385, + "learning_rate": 0.0006, + "loss": 6.558596134185791, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.014525193721055984, + "learning_rate": 0.0006, + "loss": 6.571803092956543, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.016113435849547386, + "learning_rate": 0.0006, + "loss": 6.568970680236816, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020724017173051834, + "learning_rate": 0.0006, + "loss": 6.5669403076171875, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.013963675126433372, + "learning_rate": 0.0006, + "loss": 6.586382865905762, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.01621115580201149, + "learning_rate": 0.0006, + "loss": 6.539546966552734, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.014925207942724228, + "learning_rate": 0.0006, + "loss": 6.531795501708984, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.015454350039362907, + "learning_rate": 0.0006, + "loss": 6.575699806213379, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.01623542234301567, + "learning_rate": 0.0006, + "loss": 6.582373142242432, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015452183783054352, + "learning_rate": 0.0006, + "loss": 6.576451301574707, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.021102432161569595, + "learning_rate": 0.0006, + "loss": 6.543055534362793, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.009827060624957085, + "learning_rate": 0.0006, + "loss": 6.526939392089844, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.013735136948525906, + "learning_rate": 0.0006, + "loss": 6.538825988769531, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02073509432375431, + "learning_rate": 0.0006, + "loss": 6.521315574645996, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.03110959194600582, + "learning_rate": 0.0006, + "loss": 6.561807632446289, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.045058365911245346, + "learning_rate": 0.0006, + "loss": 6.526515960693359, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.06359834969043732, + "learning_rate": 0.0006, + "loss": 6.554903030395508, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.06406767666339874, + "learning_rate": 0.0006, + "loss": 6.576272010803223, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.03403699770569801, + "learning_rate": 0.0006, + "loss": 6.5306243896484375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.036109618842601776, + "learning_rate": 0.0006, + "loss": 6.54176139831543, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.03478201478719711, + "learning_rate": 0.0006, + "loss": 6.530734062194824, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.03401544317603111, + "learning_rate": 0.0006, + "loss": 6.557794094085693, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.031629662960767746, + "learning_rate": 0.0006, + "loss": 6.564887046813965, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.02305024117231369, + "learning_rate": 0.0006, + "loss": 6.547412872314453, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030852800235152245, + "learning_rate": 0.0006, + "loss": 6.495370388031006, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.03368550166487694, + "learning_rate": 0.0006, + "loss": 6.495980262756348, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.020976359024643898, + "learning_rate": 0.0006, + "loss": 6.509410858154297, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.019652143120765686, + "learning_rate": 0.0006, + "loss": 6.493232727050781, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.024320121854543686, + "learning_rate": 0.0006, + "loss": 6.521531581878662, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.016611427068710327, + "learning_rate": 0.0006, + "loss": 6.507271766662598, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01635856367647648, + "learning_rate": 0.0006, + "loss": 6.492038726806641, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.017642095685005188, + "learning_rate": 0.0006, + "loss": 6.500349998474121, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.01679178513586521, + "learning_rate": 0.0006, + "loss": 6.500361442565918, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.011776307597756386, + "learning_rate": 0.0006, + "loss": 6.482490539550781, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.0176856629550457, + "learning_rate": 0.0006, + "loss": 6.48488712310791, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.015193155966699123, + "learning_rate": 0.0006, + "loss": 6.459545612335205, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.018623212352395058, + "learning_rate": 0.0006, + "loss": 6.478870391845703, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.012841006740927696, + "learning_rate": 0.0006, + "loss": 6.474418640136719, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.015540511347353458, + "learning_rate": 0.0006, + "loss": 6.4706525802612305, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.022695740684866905, + "learning_rate": 0.0006, + "loss": 6.455738067626953, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.026930129155516624, + "learning_rate": 0.0006, + "loss": 6.458016872406006, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.03555797412991524, + "learning_rate": 0.0006, + "loss": 6.480044364929199, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.05024682730436325, + "learning_rate": 0.0006, + "loss": 6.454564571380615, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.06616255640983582, + "learning_rate": 0.0006, + "loss": 6.433927536010742, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.059065770357847214, + "learning_rate": 0.0006, + "loss": 6.508362770080566, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.026689983904361725, + "learning_rate": 0.0006, + "loss": 6.464297294616699, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.04616398364305496, + "learning_rate": 0.0006, + "loss": 6.45448112487793, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.04980356618762016, + "learning_rate": 0.0006, + "loss": 6.4696574211120605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.0551983006298542, + "learning_rate": 0.0006, + "loss": 6.419000625610352, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.06988243758678436, + "learning_rate": 0.0006, + "loss": 6.442890167236328, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.06040226295590401, + "learning_rate": 0.0006, + "loss": 6.513258934020996, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04635024070739746, + "learning_rate": 0.0006, + "loss": 6.4710283279418945, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.04762702435255051, + "learning_rate": 0.0006, + "loss": 6.469327926635742, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.04857536032795906, + "learning_rate": 0.0006, + "loss": 6.426105499267578, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.04412693530321121, + "learning_rate": 0.0006, + "loss": 6.43179988861084, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026129672303795815, + "learning_rate": 0.0006, + "loss": 6.437918663024902, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 7.219249248504639, + "eval_runtime": 58.6837, + "eval_samples_per_second": 41.613, + "eval_steps_per_second": 1.312, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03329874947667122, + "learning_rate": 0.0006, + "loss": 6.438179969787598, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03723187372088432, + "learning_rate": 0.0006, + "loss": 6.394242286682129, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.014625726267695427, + "learning_rate": 0.0006, + "loss": 6.412118911743164, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.02917475439608097, + "learning_rate": 0.0006, + "loss": 6.453164577484131, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01811208389699459, + "learning_rate": 0.0006, + "loss": 6.396003246307373, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.027235351502895355, + "learning_rate": 0.0006, + "loss": 6.402976036071777, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.02835794910788536, + "learning_rate": 0.0006, + "loss": 6.380568504333496, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.020868154242634773, + "learning_rate": 0.0006, + "loss": 6.38240909576416, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.02113114856183529, + "learning_rate": 0.0006, + "loss": 6.377391815185547, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.018894119188189507, + "learning_rate": 0.0006, + "loss": 6.354569435119629, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.01820225641131401, + "learning_rate": 0.0006, + "loss": 6.392882347106934, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.017510728910565376, + "learning_rate": 0.0006, + "loss": 6.364162445068359, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.017893686890602112, + "learning_rate": 0.0006, + "loss": 6.373456954956055, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.013382102362811565, + "learning_rate": 0.0006, + "loss": 6.3714399337768555, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.022540833801031113, + "learning_rate": 0.0006, + "loss": 6.348170757293701, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.02110104262828827, + "learning_rate": 0.0006, + "loss": 6.413206100463867, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.014498349279165268, + "learning_rate": 0.0006, + "loss": 6.361384391784668, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.01819637045264244, + "learning_rate": 0.0006, + "loss": 6.315275192260742, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.0195972491055727, + "learning_rate": 0.0006, + "loss": 6.321994781494141, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.018822453916072845, + "learning_rate": 0.0006, + "loss": 6.345219612121582, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.019247712567448616, + "learning_rate": 0.0006, + "loss": 6.348073959350586, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.03175375238060951, + "learning_rate": 0.0006, + "loss": 6.300344944000244, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.06326444447040558, + "learning_rate": 0.0006, + "loss": 6.392579078674316, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.09667232632637024, + "learning_rate": 0.0006, + "loss": 6.36024284362793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.05774446204304695, + "learning_rate": 0.0006, + "loss": 6.365338325500488, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.08793356269598007, + "learning_rate": 0.0006, + "loss": 6.43284273147583, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.0640595331788063, + "learning_rate": 0.0006, + "loss": 6.342428207397461, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.0447838194668293, + "learning_rate": 0.0006, + "loss": 6.367971897125244, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.04768967255949974, + "learning_rate": 0.0006, + "loss": 6.35027551651001, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.027981851249933243, + "learning_rate": 0.0006, + "loss": 6.33723258972168, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.03479623422026634, + "learning_rate": 0.0006, + "loss": 6.36590051651001, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.020846687257289886, + "learning_rate": 0.0006, + "loss": 6.315143585205078, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.0312928780913353, + "learning_rate": 0.0006, + "loss": 6.340150833129883, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.020593881607055664, + "learning_rate": 0.0006, + "loss": 6.348175048828125, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.022010236978530884, + "learning_rate": 0.0006, + "loss": 6.318607330322266, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.018427059054374695, + "learning_rate": 0.0006, + "loss": 6.331822395324707, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.019964968785643578, + "learning_rate": 0.0006, + "loss": 6.287320137023926, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.02130270190536976, + "learning_rate": 0.0006, + "loss": 6.353344917297363, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.01616514101624489, + "learning_rate": 0.0006, + "loss": 6.314052581787109, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.016639474779367447, + "learning_rate": 0.0006, + "loss": 6.320351600646973, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.017620692029595375, + "learning_rate": 0.0006, + "loss": 6.311459064483643, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.013542844913899899, + "learning_rate": 0.0006, + "loss": 6.324368476867676, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01308678649365902, + "learning_rate": 0.0006, + "loss": 6.307419776916504, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.014672392047941685, + "learning_rate": 0.0006, + "loss": 6.2785420417785645, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01399585697799921, + "learning_rate": 0.0006, + "loss": 6.258619785308838, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.01330904383212328, + "learning_rate": 0.0006, + "loss": 6.299595832824707, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016613377258181572, + "learning_rate": 0.0006, + "loss": 6.2900309562683105, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.015076331794261932, + "learning_rate": 0.0006, + "loss": 6.231028079986572, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.01589728519320488, + "learning_rate": 0.0006, + "loss": 6.257922172546387, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.030972328037023544, + "learning_rate": 0.0006, + "loss": 6.285238265991211, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.05048783868551254, + "learning_rate": 0.0006, + "loss": 6.287471771240234, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.07390714436769485, + "learning_rate": 0.0006, + "loss": 6.284293174743652, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.07166943699121475, + "learning_rate": 0.0006, + "loss": 6.290704727172852, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.028792452067136765, + "learning_rate": 0.0006, + "loss": 6.243415832519531, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.04822041466832161, + "learning_rate": 0.0006, + "loss": 6.278229713439941, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.036688923835754395, + "learning_rate": 0.0006, + "loss": 6.275590896606445, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.03679114952683449, + "learning_rate": 0.0006, + "loss": 6.236509323120117, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.023249102756381035, + "learning_rate": 0.0006, + "loss": 6.230788230895996, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.02682286500930786, + "learning_rate": 0.0006, + "loss": 6.278890609741211, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.02202044241130352, + "learning_rate": 0.0006, + "loss": 6.234531402587891, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.02545459009706974, + "learning_rate": 0.0006, + "loss": 6.271790981292725, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.02012932486832142, + "learning_rate": 0.0006, + "loss": 6.247462749481201, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.02216474711894989, + "learning_rate": 0.0006, + "loss": 6.26396369934082, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.016078704968094826, + "learning_rate": 0.0006, + "loss": 6.228609085083008, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.023021478205919266, + "learning_rate": 0.0006, + "loss": 6.197244167327881, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.024766333401203156, + "learning_rate": 0.0006, + "loss": 6.181090354919434, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.025581466034054756, + "learning_rate": 0.0006, + "loss": 6.201817512512207, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.016964981332421303, + "learning_rate": 0.0006, + "loss": 6.201486587524414, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.023957721889019012, + "learning_rate": 0.0006, + "loss": 6.185701847076416, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.029775245115160942, + "learning_rate": 0.0006, + "loss": 6.242060661315918, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04007191210985184, + "learning_rate": 0.0006, + "loss": 6.19931697845459, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04882276430726051, + "learning_rate": 0.0006, + "loss": 6.1880598068237305, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 7.018043518066406, + "eval_runtime": 59.8815, + "eval_samples_per_second": 40.781, + "eval_steps_per_second": 1.286, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.07144106924533844, + "learning_rate": 0.0006, + "loss": 6.172799110412598, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.06712841987609863, + "learning_rate": 0.0006, + "loss": 6.228682518005371, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.04405239224433899, + "learning_rate": 0.0006, + "loss": 6.1827898025512695, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.03418995440006256, + "learning_rate": 0.0006, + "loss": 6.200876235961914, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.04002382606267929, + "learning_rate": 0.0006, + "loss": 6.1799211502075195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.029748762026429176, + "learning_rate": 0.0006, + "loss": 6.125422477722168, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03408357501029968, + "learning_rate": 0.0006, + "loss": 6.203458309173584, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.02227279357612133, + "learning_rate": 0.0006, + "loss": 6.135439872741699, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02727862447500229, + "learning_rate": 0.0006, + "loss": 6.202380180358887, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.026019075885415077, + "learning_rate": 0.0006, + "loss": 6.201343536376953, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.027227282524108887, + "learning_rate": 0.0006, + "loss": 6.164059638977051, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.027649980038404465, + "learning_rate": 0.0006, + "loss": 6.135448932647705, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02168039232492447, + "learning_rate": 0.0006, + "loss": 6.154831886291504, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02101467363536358, + "learning_rate": 0.0006, + "loss": 6.128695487976074, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.018344245851039886, + "learning_rate": 0.0006, + "loss": 6.136524200439453, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020323453471064568, + "learning_rate": 0.0006, + "loss": 6.1206488609313965, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.01750849559903145, + "learning_rate": 0.0006, + "loss": 6.128170967102051, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017778221517801285, + "learning_rate": 0.0006, + "loss": 6.150458812713623, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.01863938942551613, + "learning_rate": 0.0006, + "loss": 6.100106716156006, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.020196080207824707, + "learning_rate": 0.0006, + "loss": 6.149346351623535, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.03203890845179558, + "learning_rate": 0.0006, + "loss": 6.124648094177246, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.05325170233845711, + "learning_rate": 0.0006, + "loss": 6.115618705749512, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.05888345465064049, + "learning_rate": 0.0006, + "loss": 6.165221214294434, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.026873115450143814, + "learning_rate": 0.0006, + "loss": 6.039590835571289, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.06304649263620377, + "learning_rate": 0.0006, + "loss": 6.131900787353516, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.08567557483911514, + "learning_rate": 0.0006, + "loss": 6.218114852905273, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.08868218213319778, + "learning_rate": 0.0006, + "loss": 6.216706275939941, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.03955601155757904, + "learning_rate": 0.0006, + "loss": 6.1566925048828125, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.04979498311877251, + "learning_rate": 0.0006, + "loss": 6.19327449798584, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.03253026679158211, + "learning_rate": 0.0006, + "loss": 6.133993148803711, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03016652725636959, + "learning_rate": 0.0006, + "loss": 6.174962997436523, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.03159862384200096, + "learning_rate": 0.0006, + "loss": 6.082124710083008, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.027673959732055664, + "learning_rate": 0.0006, + "loss": 6.111873149871826, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.022789714857935905, + "learning_rate": 0.0006, + "loss": 6.129437446594238, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.030061012133955956, + "learning_rate": 0.0006, + "loss": 6.126370429992676, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.019007068127393723, + "learning_rate": 0.0006, + "loss": 6.07887077331543, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.01897653006017208, + "learning_rate": 0.0006, + "loss": 6.096936225891113, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.024659698829054832, + "learning_rate": 0.0006, + "loss": 6.0933380126953125, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.021924549713730812, + "learning_rate": 0.0006, + "loss": 6.105311870574951, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.018817558884620667, + "learning_rate": 0.0006, + "loss": 6.058895111083984, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.02477840520441532, + "learning_rate": 0.0006, + "loss": 6.038928031921387, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.018968632444739342, + "learning_rate": 0.0006, + "loss": 6.1368937492370605, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.02409801445901394, + "learning_rate": 0.0006, + "loss": 6.110324859619141, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.017595326527953148, + "learning_rate": 0.0006, + "loss": 6.074368476867676, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.01875406503677368, + "learning_rate": 0.0006, + "loss": 6.121023654937744, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.018781628459692, + "learning_rate": 0.0006, + "loss": 6.064914703369141, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.017003126442432404, + "learning_rate": 0.0006, + "loss": 6.039608001708984, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.017665037885308266, + "learning_rate": 0.0006, + "loss": 6.053671836853027, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.02066265605390072, + "learning_rate": 0.0006, + "loss": 6.090536117553711, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02888769656419754, + "learning_rate": 0.0006, + "loss": 5.9858903884887695, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.03744485229253769, + "learning_rate": 0.0006, + "loss": 6.086264610290527, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04719431325793266, + "learning_rate": 0.0006, + "loss": 6.09747314453125, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.044963907450437546, + "learning_rate": 0.0006, + "loss": 6.091272354125977, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03810194507241249, + "learning_rate": 0.0006, + "loss": 6.084904193878174, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.04626769199967384, + "learning_rate": 0.0006, + "loss": 6.017384052276611, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.03459552302956581, + "learning_rate": 0.0006, + "loss": 6.111314296722412, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.027385666966438293, + "learning_rate": 0.0006, + "loss": 6.002509117126465, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.03417652100324631, + "learning_rate": 0.0006, + "loss": 6.0421061515808105, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.04415450245141983, + "learning_rate": 0.0006, + "loss": 6.035305023193359, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.03259635716676712, + "learning_rate": 0.0006, + "loss": 6.025840759277344, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.024941129609942436, + "learning_rate": 0.0006, + "loss": 6.043416976928711, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.02434500865638256, + "learning_rate": 0.0006, + "loss": 6.002052307128906, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.01876584254205227, + "learning_rate": 0.0006, + "loss": 6.045035362243652, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.02254367619752884, + "learning_rate": 0.0006, + "loss": 5.972503662109375, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.024504603818058968, + "learning_rate": 0.0006, + "loss": 5.992846488952637, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.02317827194929123, + "learning_rate": 0.0006, + "loss": 6.007697582244873, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02225821278989315, + "learning_rate": 0.0006, + "loss": 6.01463508605957, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.01811717264354229, + "learning_rate": 0.0006, + "loss": 5.995022773742676, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.019589431583881378, + "learning_rate": 0.0006, + "loss": 5.990633964538574, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015422016382217407, + "learning_rate": 0.0006, + "loss": 6.011263847351074, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.9954938888549805, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.022172044962644577, + "learning_rate": 0.0006, + "loss": 5.95679235458374, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 6.75809383392334, + "eval_runtime": 72.8926, + "eval_samples_per_second": 33.501, + "eval_steps_per_second": 1.056, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.029082221910357475, + "learning_rate": 0.0006, + "loss": 6.018675804138184, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.03959495946764946, + "learning_rate": 0.0006, + "loss": 5.93310546875, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.043756261467933655, + "learning_rate": 0.0006, + "loss": 5.965944290161133, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.04996631667017937, + "learning_rate": 0.0006, + "loss": 6.0381059646606445, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.05721999332308769, + "learning_rate": 0.0006, + "loss": 5.968381881713867, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.07594525814056396, + "learning_rate": 0.0006, + "loss": 6.027225494384766, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.06861685961484909, + "learning_rate": 0.0006, + "loss": 6.035175323486328, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04156472906470299, + "learning_rate": 0.0006, + "loss": 5.9500017166137695, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.045399412512779236, + "learning_rate": 0.0006, + "loss": 5.9953389167785645, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.04399585723876953, + "learning_rate": 0.0006, + "loss": 5.986051559448242, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05764509737491608, + "learning_rate": 0.0006, + "loss": 6.005392074584961, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.060871850699186325, + "learning_rate": 0.0006, + "loss": 6.01304817199707, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.03381241112947464, + "learning_rate": 0.0006, + "loss": 5.977019309997559, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.03384104371070862, + "learning_rate": 0.0006, + "loss": 5.988002777099609, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.03120286948978901, + "learning_rate": 0.0006, + "loss": 5.968227386474609, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.026532653719186783, + "learning_rate": 0.0006, + "loss": 5.936810493469238, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.024956922978162766, + "learning_rate": 0.0006, + "loss": 5.9348015785217285, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.025926144793629646, + "learning_rate": 0.0006, + "loss": 5.952630519866943, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.02893793396651745, + "learning_rate": 0.0006, + "loss": 5.9569854736328125, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.0320001058280468, + "learning_rate": 0.0006, + "loss": 5.917050838470459, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.03925463184714317, + "learning_rate": 0.0006, + "loss": 5.9925127029418945, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.05096640810370445, + "learning_rate": 0.0006, + "loss": 5.926358222961426, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.041637953370809555, + "learning_rate": 0.0006, + "loss": 5.954983711242676, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03138125315308571, + "learning_rate": 0.0006, + "loss": 5.893692970275879, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.029299303889274597, + "learning_rate": 0.0006, + "loss": 5.94670295715332, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.023969192057847977, + "learning_rate": 0.0006, + "loss": 5.925658702850342, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.025406932458281517, + "learning_rate": 0.0006, + "loss": 5.966875076293945, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.024309059605002403, + "learning_rate": 0.0006, + "loss": 5.932581424713135, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.017237074673175812, + "learning_rate": 0.0006, + "loss": 5.898823261260986, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.025594931095838547, + "learning_rate": 0.0006, + "loss": 5.9285993576049805, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.025184160098433495, + "learning_rate": 0.0006, + "loss": 5.915966510772705, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019004398956894875, + "learning_rate": 0.0006, + "loss": 5.893395900726318, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.021802261471748352, + "learning_rate": 0.0006, + "loss": 5.953311443328857, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02202655002474785, + "learning_rate": 0.0006, + "loss": 5.953000545501709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.02067241445183754, + "learning_rate": 0.0006, + "loss": 5.869247913360596, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.015430927276611328, + "learning_rate": 0.0006, + "loss": 5.855883598327637, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.019917506724596024, + "learning_rate": 0.0006, + "loss": 5.9284539222717285, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.024291323497891426, + "learning_rate": 0.0006, + "loss": 5.973604202270508, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.022133296355605125, + "learning_rate": 0.0006, + "loss": 5.85693359375, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0254366435110569, + "learning_rate": 0.0006, + "loss": 5.939558982849121, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.030134694650769234, + "learning_rate": 0.0006, + "loss": 5.892992973327637, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.03845192864537239, + "learning_rate": 0.0006, + "loss": 5.89303731918335, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.042409464716911316, + "learning_rate": 0.0006, + "loss": 5.914068222045898, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.0454271174967289, + "learning_rate": 0.0006, + "loss": 5.853579044342041, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.04934931918978691, + "learning_rate": 0.0006, + "loss": 5.881861686706543, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.0519695058465004, + "learning_rate": 0.0006, + "loss": 5.872620582580566, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.06172563508152962, + "learning_rate": 0.0006, + "loss": 5.938283443450928, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.08416368812322617, + "learning_rate": 0.0006, + "loss": 5.943403720855713, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.07311230152845383, + "learning_rate": 0.0006, + "loss": 5.954029083251953, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.09851036965847015, + "learning_rate": 0.0006, + "loss": 5.992746829986572, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.07521767914295197, + "learning_rate": 0.0006, + "loss": 5.956884384155273, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.05720978602766991, + "learning_rate": 0.0006, + "loss": 5.911205291748047, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.05360936373472214, + "learning_rate": 0.0006, + "loss": 5.9734787940979, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.035826463252305984, + "learning_rate": 0.0006, + "loss": 5.95751953125, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.030137093737721443, + "learning_rate": 0.0006, + "loss": 5.91911506652832, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030039150267839432, + "learning_rate": 0.0006, + "loss": 5.910290718078613, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.03349437564611435, + "learning_rate": 0.0006, + "loss": 5.904623985290527, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02211189828813076, + "learning_rate": 0.0006, + "loss": 5.909668445587158, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.022601254284381866, + "learning_rate": 0.0006, + "loss": 5.942885875701904, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.02197270467877388, + "learning_rate": 0.0006, + "loss": 5.867103576660156, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.022765018045902252, + "learning_rate": 0.0006, + "loss": 5.894528865814209, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021181151270866394, + "learning_rate": 0.0006, + "loss": 5.896910667419434, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.01878993771970272, + "learning_rate": 0.0006, + "loss": 5.860502243041992, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.014679803512990475, + "learning_rate": 0.0006, + "loss": 5.8725266456604, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.017919695004820824, + "learning_rate": 0.0006, + "loss": 5.88902473449707, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.01437009871006012, + "learning_rate": 0.0006, + "loss": 5.868186950683594, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.014293290674686432, + "learning_rate": 0.0006, + "loss": 5.833364963531494, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.015440084971487522, + "learning_rate": 0.0006, + "loss": 5.860260009765625, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.014292514882981777, + "learning_rate": 0.0006, + "loss": 5.844939231872559, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.015702251344919205, + "learning_rate": 0.0006, + "loss": 5.825929164886475, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.013372798450291157, + "learning_rate": 0.0006, + "loss": 5.83062744140625, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.013794616796076298, + "learning_rate": 0.0006, + "loss": 5.7897820472717285, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 6.600830078125, + "eval_runtime": 71.4316, + "eval_samples_per_second": 34.187, + "eval_steps_per_second": 1.078, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.01262663770467043, + "learning_rate": 0.0006, + "loss": 5.802824020385742, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.016375157982110977, + "learning_rate": 0.0006, + "loss": 5.8018317222595215, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.010541851632297039, + "learning_rate": 0.0006, + "loss": 5.848208427429199, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.011726257391273975, + "learning_rate": 0.0006, + "loss": 5.771823883056641, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01124380249530077, + "learning_rate": 0.0006, + "loss": 5.773855209350586, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.010801885277032852, + "learning_rate": 0.0006, + "loss": 5.847390651702881, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.013018375262618065, + "learning_rate": 0.0006, + "loss": 5.786786079406738, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.014233003370463848, + "learning_rate": 0.0006, + "loss": 5.7913713455200195, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.01465831883251667, + "learning_rate": 0.0006, + "loss": 5.821170330047607, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.017383430153131485, + "learning_rate": 0.0006, + "loss": 5.811469078063965, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.020328007638454437, + "learning_rate": 0.0006, + "loss": 5.718976020812988, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.01776740700006485, + "learning_rate": 0.0006, + "loss": 5.80593729019165, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.017713800072669983, + "learning_rate": 0.0006, + "loss": 5.744930267333984, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.028774067759513855, + "learning_rate": 0.0006, + "loss": 5.85262393951416, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.05153049901127815, + "learning_rate": 0.0006, + "loss": 5.739948272705078, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.06882233172655106, + "learning_rate": 0.0006, + "loss": 5.844270706176758, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.0420759879052639, + "learning_rate": 0.0006, + "loss": 5.79060173034668, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.05598323419690132, + "learning_rate": 0.0006, + "loss": 5.833735466003418, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.05011717975139618, + "learning_rate": 0.0006, + "loss": 5.840887069702148, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.040782492607831955, + "learning_rate": 0.0006, + "loss": 5.7905449867248535, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.030804185196757317, + "learning_rate": 0.0006, + "loss": 5.802691459655762, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.032001759856939316, + "learning_rate": 0.0006, + "loss": 5.726612091064453, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.0295492485165596, + "learning_rate": 0.0006, + "loss": 5.798826694488525, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.034406546503305435, + "learning_rate": 0.0006, + "loss": 5.720219135284424, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.02627669647336006, + "learning_rate": 0.0006, + "loss": 5.723702430725098, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.0282363910228014, + "learning_rate": 0.0006, + "loss": 5.818911552429199, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.017894433811306953, + "learning_rate": 0.0006, + "loss": 5.772168159484863, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.019017303362488747, + "learning_rate": 0.0006, + "loss": 5.774587631225586, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.022159283980727196, + "learning_rate": 0.0006, + "loss": 5.812951564788818, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02198116108775139, + "learning_rate": 0.0006, + "loss": 5.808734893798828, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.017741188406944275, + "learning_rate": 0.0006, + "loss": 5.776142120361328, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.017197400331497192, + "learning_rate": 0.0006, + "loss": 5.7471208572387695, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.01752566732466221, + "learning_rate": 0.0006, + "loss": 5.780961990356445, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.017961280420422554, + "learning_rate": 0.0006, + "loss": 5.713686943054199, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.02227521874010563, + "learning_rate": 0.0006, + "loss": 5.717774868011475, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.024075504392385483, + "learning_rate": 0.0006, + "loss": 5.715015411376953, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.029195033013820648, + "learning_rate": 0.0006, + "loss": 5.659659385681152, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.02959120273590088, + "learning_rate": 0.0006, + "loss": 5.78981351852417, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.021854368969798088, + "learning_rate": 0.0006, + "loss": 5.750844955444336, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017120081931352615, + "learning_rate": 0.0006, + "loss": 5.672736167907715, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.020712822675704956, + "learning_rate": 0.0006, + "loss": 5.708046913146973, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.019186170771718025, + "learning_rate": 0.0006, + "loss": 5.71742057800293, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.01985814794898033, + "learning_rate": 0.0006, + "loss": 5.735311508178711, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.01965234987437725, + "learning_rate": 0.0006, + "loss": 5.728727340698242, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.019827688112854958, + "learning_rate": 0.0006, + "loss": 5.743068695068359, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.025500580668449402, + "learning_rate": 0.0006, + "loss": 5.76585578918457, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.03701791539788246, + "learning_rate": 0.0006, + "loss": 5.728367805480957, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.04173659533262253, + "learning_rate": 0.0006, + "loss": 5.683089256286621, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.031390558928251266, + "learning_rate": 0.0006, + "loss": 5.740921974182129, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.03970158472657204, + "learning_rate": 0.0006, + "loss": 5.730632781982422, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.03865141049027443, + "learning_rate": 0.0006, + "loss": 5.686657905578613, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.04856107011437416, + "learning_rate": 0.0006, + "loss": 5.657513618469238, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.04016133397817612, + "learning_rate": 0.0006, + "loss": 5.754552841186523, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.03961458429694176, + "learning_rate": 0.0006, + "loss": 5.672088623046875, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.03403536602854729, + "learning_rate": 0.0006, + "loss": 5.736673355102539, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03766333684325218, + "learning_rate": 0.0006, + "loss": 5.716219902038574, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.04085501655936241, + "learning_rate": 0.0006, + "loss": 5.678958415985107, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.036888182163238525, + "learning_rate": 0.0006, + "loss": 5.6990461349487305, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.0492791011929512, + "learning_rate": 0.0006, + "loss": 5.78053092956543, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.05572429671883583, + "learning_rate": 0.0006, + "loss": 5.7543535232543945, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.047000352293252945, + "learning_rate": 0.0006, + "loss": 5.7174272537231445, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.058798179030418396, + "learning_rate": 0.0006, + "loss": 5.793713569641113, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.046992287039756775, + "learning_rate": 0.0006, + "loss": 5.784457683563232, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.04382370039820671, + "learning_rate": 0.0006, + "loss": 5.793731212615967, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.040616173297166824, + "learning_rate": 0.0006, + "loss": 5.69599723815918, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03446170687675476, + "learning_rate": 0.0006, + "loss": 5.694674968719482, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029936550185084343, + "learning_rate": 0.0006, + "loss": 5.6704487800598145, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02224753610789776, + "learning_rate": 0.0006, + "loss": 5.728894233703613, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.024346651509404182, + "learning_rate": 0.0006, + "loss": 5.772412300109863, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.022997068241238594, + "learning_rate": 0.0006, + "loss": 5.645310401916504, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.024918686598539352, + "learning_rate": 0.0006, + "loss": 5.671009063720703, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.020067067816853523, + "learning_rate": 0.0006, + "loss": 5.637441635131836, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 6.387807846069336, + "eval_runtime": 72.3372, + "eval_samples_per_second": 33.759, + "eval_steps_per_second": 1.064, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02015046589076519, + "learning_rate": 0.0006, + "loss": 5.598219871520996, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.017398346215486526, + "learning_rate": 0.0006, + "loss": 5.66834831237793, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.016871167346835136, + "learning_rate": 0.0006, + "loss": 5.673916816711426, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.01695060543715954, + "learning_rate": 0.0006, + "loss": 5.646677017211914, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.019298404455184937, + "learning_rate": 0.0006, + "loss": 5.663064479827881, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018699361011385918, + "learning_rate": 0.0006, + "loss": 5.615242958068848, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.019567854702472687, + "learning_rate": 0.0006, + "loss": 5.5521240234375, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.018347466364502907, + "learning_rate": 0.0006, + "loss": 5.5929107666015625, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.020350053906440735, + "learning_rate": 0.0006, + "loss": 5.6380109786987305, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.016790535300970078, + "learning_rate": 0.0006, + "loss": 5.62994384765625, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.01413649506866932, + "learning_rate": 0.0006, + "loss": 5.6301727294921875, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.013762339949607849, + "learning_rate": 0.0006, + "loss": 5.584138870239258, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.014790690504014492, + "learning_rate": 0.0006, + "loss": 5.666123390197754, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.015471262857317924, + "learning_rate": 0.0006, + "loss": 5.587042808532715, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.015804501250386238, + "learning_rate": 0.0006, + "loss": 5.627497673034668, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.017627960070967674, + "learning_rate": 0.0006, + "loss": 5.556890487670898, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.02074616774916649, + "learning_rate": 0.0006, + "loss": 5.586511135101318, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.026389935985207558, + "learning_rate": 0.0006, + "loss": 5.598520278930664, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03736288473010063, + "learning_rate": 0.0006, + "loss": 5.575788497924805, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.03523678705096245, + "learning_rate": 0.0006, + "loss": 5.586930751800537, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.024266984313726425, + "learning_rate": 0.0006, + "loss": 5.596694469451904, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.033640433102846146, + "learning_rate": 0.0006, + "loss": 5.536456108093262, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.04689764603972435, + "learning_rate": 0.0006, + "loss": 5.61761999130249, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.04910685867071152, + "learning_rate": 0.0006, + "loss": 5.6298418045043945, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.03152550011873245, + "learning_rate": 0.0006, + "loss": 5.639697074890137, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.037753574550151825, + "learning_rate": 0.0006, + "loss": 5.697381019592285, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.04877670109272003, + "learning_rate": 0.0006, + "loss": 5.605184555053711, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.04228494316339493, + "learning_rate": 0.0006, + "loss": 5.624238967895508, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.03575790300965309, + "learning_rate": 0.0006, + "loss": 5.6037983894348145, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.036425549536943436, + "learning_rate": 0.0006, + "loss": 5.62905216217041, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.04001355543732643, + "learning_rate": 0.0006, + "loss": 5.581656455993652, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.05040939524769783, + "learning_rate": 0.0006, + "loss": 5.621358871459961, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.04140249267220497, + "learning_rate": 0.0006, + "loss": 5.668188095092773, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.033116403967142105, + "learning_rate": 0.0006, + "loss": 5.580477714538574, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.03864264488220215, + "learning_rate": 0.0006, + "loss": 5.539153099060059, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.03026588261127472, + "learning_rate": 0.0006, + "loss": 5.619814872741699, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.02381783165037632, + "learning_rate": 0.0006, + "loss": 5.64422607421875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02427830547094345, + "learning_rate": 0.0006, + "loss": 5.659873008728027, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.021709900349378586, + "learning_rate": 0.0006, + "loss": 5.590245723724365, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.022812914103269577, + "learning_rate": 0.0006, + "loss": 5.610434532165527, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.019250890240073204, + "learning_rate": 0.0006, + "loss": 5.588078498840332, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.021995501592755318, + "learning_rate": 0.0006, + "loss": 5.541754245758057, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01477806642651558, + "learning_rate": 0.0006, + "loss": 5.549173355102539, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.016680821776390076, + "learning_rate": 0.0006, + "loss": 5.58221435546875, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013848318718373775, + "learning_rate": 0.0006, + "loss": 5.5505876541137695, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.016387000679969788, + "learning_rate": 0.0006, + "loss": 5.570460319519043, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.016496233642101288, + "learning_rate": 0.0006, + "loss": 5.513058662414551, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01675555482506752, + "learning_rate": 0.0006, + "loss": 5.559998989105225, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.016939565539360046, + "learning_rate": 0.0006, + "loss": 5.56727409362793, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.01679506152868271, + "learning_rate": 0.0006, + "loss": 5.603643417358398, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01656349189579487, + "learning_rate": 0.0006, + "loss": 5.54292631149292, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.01975332945585251, + "learning_rate": 0.0006, + "loss": 5.5225677490234375, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.02329389750957489, + "learning_rate": 0.0006, + "loss": 5.537827491760254, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.026616891846060753, + "learning_rate": 0.0006, + "loss": 5.500011920928955, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.025511575862765312, + "learning_rate": 0.0006, + "loss": 5.5316619873046875, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.025009512901306152, + "learning_rate": 0.0006, + "loss": 5.468493938446045, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02946239896118641, + "learning_rate": 0.0006, + "loss": 5.5116753578186035, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.036607250571250916, + "learning_rate": 0.0006, + "loss": 5.518636226654053, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.04347814992070198, + "learning_rate": 0.0006, + "loss": 5.572815895080566, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.034463658928871155, + "learning_rate": 0.0006, + "loss": 5.520634651184082, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.03486703336238861, + "learning_rate": 0.0006, + "loss": 5.545441150665283, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.04476739838719368, + "learning_rate": 0.0006, + "loss": 5.554450035095215, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.048745233565568924, + "learning_rate": 0.0006, + "loss": 5.554961204528809, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.04636877775192261, + "learning_rate": 0.0006, + "loss": 5.574655055999756, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.032686877995729446, + "learning_rate": 0.0006, + "loss": 5.5734639167785645, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.03075302764773369, + "learning_rate": 0.0006, + "loss": 5.487662315368652, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.028860175982117653, + "learning_rate": 0.0006, + "loss": 5.519798755645752, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.028599515557289124, + "learning_rate": 0.0006, + "loss": 5.558349609375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.02384260669350624, + "learning_rate": 0.0006, + "loss": 5.520661354064941, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.023180771619081497, + "learning_rate": 0.0006, + "loss": 5.496896743774414, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.021167006343603134, + "learning_rate": 0.0006, + "loss": 5.517192363739014, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02069013938307762, + "learning_rate": 0.0006, + "loss": 5.447656154632568, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 6.161121845245361, + "eval_runtime": 71.5972, + "eval_samples_per_second": 34.107, + "eval_steps_per_second": 1.075, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.022983292117714882, + "learning_rate": 0.0006, + "loss": 5.441346645355225, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.022833511233329773, + "learning_rate": 0.0006, + "loss": 5.45123815536499, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.020015383139252663, + "learning_rate": 0.0006, + "loss": 5.430389404296875, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.016957225278019905, + "learning_rate": 0.0006, + "loss": 5.473883628845215, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.02090892568230629, + "learning_rate": 0.0006, + "loss": 5.467441558837891, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.022806301712989807, + "learning_rate": 0.0006, + "loss": 5.5453081130981445, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.022816946730017662, + "learning_rate": 0.0006, + "loss": 5.426213264465332, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.020761270076036453, + "learning_rate": 0.0006, + "loss": 5.464138507843018, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01982618309557438, + "learning_rate": 0.0006, + "loss": 5.428033828735352, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.021953724324703217, + "learning_rate": 0.0006, + "loss": 5.456646919250488, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01871841959655285, + "learning_rate": 0.0006, + "loss": 5.508403778076172, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.01810353808104992, + "learning_rate": 0.0006, + "loss": 5.442957878112793, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.017609186470508575, + "learning_rate": 0.0006, + "loss": 5.4774699211120605, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.01931103691458702, + "learning_rate": 0.0006, + "loss": 5.429748058319092, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.019401391968131065, + "learning_rate": 0.0006, + "loss": 5.466834545135498, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.018862580880522728, + "learning_rate": 0.0006, + "loss": 5.431701183319092, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.01796964928507805, + "learning_rate": 0.0006, + "loss": 5.388613700866699, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.021166296675801277, + "learning_rate": 0.0006, + "loss": 5.454118251800537, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.028568729758262634, + "learning_rate": 0.0006, + "loss": 5.510260581970215, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.03902891278266907, + "learning_rate": 0.0006, + "loss": 5.386396408081055, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.039114370942115784, + "learning_rate": 0.0006, + "loss": 5.457526206970215, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.03560924902558327, + "learning_rate": 0.0006, + "loss": 5.4149980545043945, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.03745810315012932, + "learning_rate": 0.0006, + "loss": 5.464084625244141, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.03383171558380127, + "learning_rate": 0.0006, + "loss": 5.45137882232666, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.027873128652572632, + "learning_rate": 0.0006, + "loss": 5.485784530639648, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.02697703056037426, + "learning_rate": 0.0006, + "loss": 5.459152698516846, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.02646288089454174, + "learning_rate": 0.0006, + "loss": 5.433281421661377, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.02655179798603058, + "learning_rate": 0.0006, + "loss": 5.456627368927002, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02633797563612461, + "learning_rate": 0.0006, + "loss": 5.484803676605225, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.031187085434794426, + "learning_rate": 0.0006, + "loss": 5.362293243408203, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02971349097788334, + "learning_rate": 0.0006, + "loss": 5.4369611740112305, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.02996021695435047, + "learning_rate": 0.0006, + "loss": 5.419068813323975, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.032519739121198654, + "learning_rate": 0.0006, + "loss": 5.422189712524414, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.03403490409255028, + "learning_rate": 0.0006, + "loss": 5.470074653625488, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.036989886313676834, + "learning_rate": 0.0006, + "loss": 5.460174560546875, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.03427419438958168, + "learning_rate": 0.0006, + "loss": 5.495920658111572, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.02519620768725872, + "learning_rate": 0.0006, + "loss": 5.474429130554199, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.02473096363246441, + "learning_rate": 0.0006, + "loss": 5.395847320556641, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.02792714536190033, + "learning_rate": 0.0006, + "loss": 5.376262664794922, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.024172596633434296, + "learning_rate": 0.0006, + "loss": 5.459731101989746, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02304821088910103, + "learning_rate": 0.0006, + "loss": 5.479475498199463, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.02342340536415577, + "learning_rate": 0.0006, + "loss": 5.387955665588379, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021962853148579597, + "learning_rate": 0.0006, + "loss": 5.385697364807129, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.019193042069673538, + "learning_rate": 0.0006, + "loss": 5.431601047515869, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.021378040313720703, + "learning_rate": 0.0006, + "loss": 5.3070969581604, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.021641355007886887, + "learning_rate": 0.0006, + "loss": 5.360584735870361, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.019899170845746994, + "learning_rate": 0.0006, + "loss": 5.398155212402344, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.02298869751393795, + "learning_rate": 0.0006, + "loss": 5.410561561584473, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.02815551869571209, + "learning_rate": 0.0006, + "loss": 5.298604965209961, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.03041219152510166, + "learning_rate": 0.0006, + "loss": 5.425869941711426, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.025710400193929672, + "learning_rate": 0.0006, + "loss": 5.344316482543945, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.018438996747136116, + "learning_rate": 0.0006, + "loss": 5.379913330078125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.016683703288435936, + "learning_rate": 0.0006, + "loss": 5.338386058807373, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.017252853140234947, + "learning_rate": 0.0006, + "loss": 5.360256195068359, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.019606001675128937, + "learning_rate": 0.0006, + "loss": 5.401415824890137, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.016299517825245857, + "learning_rate": 0.0006, + "loss": 5.39270544052124, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.016657982021570206, + "learning_rate": 0.0006, + "loss": 5.359323978424072, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.017462270334362984, + "learning_rate": 0.0006, + "loss": 5.453655242919922, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01723778434097767, + "learning_rate": 0.0006, + "loss": 5.454940319061279, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.019071755930781364, + "learning_rate": 0.0006, + "loss": 5.355914115905762, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02169782482087612, + "learning_rate": 0.0006, + "loss": 5.396806716918945, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.025049131363630295, + "learning_rate": 0.0006, + "loss": 5.347422122955322, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.024977540597319603, + "learning_rate": 0.0006, + "loss": 5.32526969909668, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.024223418906331062, + "learning_rate": 0.0006, + "loss": 5.372445106506348, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.02519913762807846, + "learning_rate": 0.0006, + "loss": 5.280451774597168, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.0255630724132061, + "learning_rate": 0.0006, + "loss": 5.462911605834961, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.022740041837096214, + "learning_rate": 0.0006, + "loss": 5.319089412689209, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02038649097084999, + "learning_rate": 0.0006, + "loss": 5.37426233291626, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.020509786903858185, + "learning_rate": 0.0006, + "loss": 5.337924957275391, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.021707357838749886, + "learning_rate": 0.0006, + "loss": 5.374853134155273, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02714136242866516, + "learning_rate": 0.0006, + "loss": 5.2982282638549805, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.027913225814700127, + "learning_rate": 0.0006, + "loss": 5.406044960021973, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.948610305786133, + "eval_runtime": 75.79, + "eval_samples_per_second": 32.221, + "eval_steps_per_second": 1.016, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.023761706426739693, + "learning_rate": 0.0006, + "loss": 5.378547191619873, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.026866815984249115, + "learning_rate": 0.0006, + "loss": 5.356947898864746, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.03335791453719139, + "learning_rate": 0.0006, + "loss": 5.264157772064209, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.03153866529464722, + "learning_rate": 0.0006, + "loss": 5.363075256347656, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.03270367905497551, + "learning_rate": 0.0006, + "loss": 5.298289775848389, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.030872846022248268, + "learning_rate": 0.0006, + "loss": 5.324321269989014, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.028438350185751915, + "learning_rate": 0.0006, + "loss": 5.3512115478515625, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.030846193432807922, + "learning_rate": 0.0006, + "loss": 5.304851531982422, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.03879433870315552, + "learning_rate": 0.0006, + "loss": 5.289634704589844, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.03665967285633087, + "learning_rate": 0.0006, + "loss": 5.352825164794922, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.03454619646072388, + "learning_rate": 0.0006, + "loss": 5.346499443054199, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.043245580047369, + "learning_rate": 0.0006, + "loss": 5.344263076782227, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.04450082778930664, + "learning_rate": 0.0006, + "loss": 5.447438716888428, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.04980820417404175, + "learning_rate": 0.0006, + "loss": 5.364802360534668, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.04757903143763542, + "learning_rate": 0.0006, + "loss": 5.363414287567139, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.048647016286849976, + "learning_rate": 0.0006, + "loss": 5.379604339599609, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.04497060924768448, + "learning_rate": 0.0006, + "loss": 5.354029178619385, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.032642051577568054, + "learning_rate": 0.0006, + "loss": 5.335702419281006, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.037766437977552414, + "learning_rate": 0.0006, + "loss": 5.331804275512695, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.030145324766635895, + "learning_rate": 0.0006, + "loss": 5.327176094055176, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.032248858362436295, + "learning_rate": 0.0006, + "loss": 5.2410712242126465, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.026990503072738647, + "learning_rate": 0.0006, + "loss": 5.369958877563477, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.023413540795445442, + "learning_rate": 0.0006, + "loss": 5.322418212890625, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.025248292833566666, + "learning_rate": 0.0006, + "loss": 5.300156116485596, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.021037699654698372, + "learning_rate": 0.0006, + "loss": 5.26978063583374, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.0155200669541955, + "learning_rate": 0.0006, + "loss": 5.297061920166016, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.017867842689156532, + "learning_rate": 0.0006, + "loss": 5.312634468078613, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.01675534062087536, + "learning_rate": 0.0006, + "loss": 5.290635108947754, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.015546303242444992, + "learning_rate": 0.0006, + "loss": 5.289699554443359, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.01612856611609459, + "learning_rate": 0.0006, + "loss": 5.312010765075684, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.015225349925458431, + "learning_rate": 0.0006, + "loss": 5.255838871002197, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015317161567509174, + "learning_rate": 0.0006, + "loss": 5.234170436859131, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.014820792712271214, + "learning_rate": 0.0006, + "loss": 5.255383014678955, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.01343776099383831, + "learning_rate": 0.0006, + "loss": 5.256904602050781, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.01418706402182579, + "learning_rate": 0.0006, + "loss": 5.272562026977539, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.013406263664364815, + "learning_rate": 0.0006, + "loss": 5.17042350769043, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.01331746019423008, + "learning_rate": 0.0006, + "loss": 5.267889022827148, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.014625098556280136, + "learning_rate": 0.0006, + "loss": 5.281431198120117, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.01346531230956316, + "learning_rate": 0.0006, + "loss": 5.286074638366699, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.01303934771567583, + "learning_rate": 0.0006, + "loss": 5.344804763793945, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.014180841855704784, + "learning_rate": 0.0006, + "loss": 5.252955436706543, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.014279021881520748, + "learning_rate": 0.0006, + "loss": 5.283501625061035, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.013922203332185745, + "learning_rate": 0.0006, + "loss": 5.225350379943848, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.014055478386580944, + "learning_rate": 0.0006, + "loss": 5.283530235290527, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.013623543083667755, + "learning_rate": 0.0006, + "loss": 5.204877853393555, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.014378432184457779, + "learning_rate": 0.0006, + "loss": 5.275735378265381, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.017884068191051483, + "learning_rate": 0.0006, + "loss": 5.256485939025879, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.0203024260699749, + "learning_rate": 0.0006, + "loss": 5.292258262634277, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.021196166053414345, + "learning_rate": 0.0006, + "loss": 5.30052375793457, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.02249213121831417, + "learning_rate": 0.0006, + "loss": 5.236924648284912, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.019980570301413536, + "learning_rate": 0.0006, + "loss": 5.290444374084473, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.022030364722013474, + "learning_rate": 0.0006, + "loss": 5.287881851196289, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.02438683621585369, + "learning_rate": 0.0006, + "loss": 5.230462074279785, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.026895932853221893, + "learning_rate": 0.0006, + "loss": 5.232667446136475, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.02861972153186798, + "learning_rate": 0.0006, + "loss": 5.205913066864014, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.033064406365156174, + "learning_rate": 0.0006, + "loss": 5.242166042327881, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.030850905925035477, + "learning_rate": 0.0006, + "loss": 5.2503581047058105, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.03139309585094452, + "learning_rate": 0.0006, + "loss": 5.222465515136719, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.035760655999183655, + "learning_rate": 0.0006, + "loss": 5.279786586761475, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.036571063101291656, + "learning_rate": 0.0006, + "loss": 5.237204551696777, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.03467971459031105, + "learning_rate": 0.0006, + "loss": 5.210154056549072, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.053154416382312775, + "learning_rate": 0.0006, + "loss": 5.295984268188477, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.05374281480908394, + "learning_rate": 0.0006, + "loss": 5.293900012969971, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.05851082503795624, + "learning_rate": 0.0006, + "loss": 5.3758931159973145, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.05145686864852905, + "learning_rate": 0.0006, + "loss": 5.309441566467285, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.04484057426452637, + "learning_rate": 0.0006, + "loss": 5.297204494476318, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.05348053574562073, + "learning_rate": 0.0006, + "loss": 5.278589248657227, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.05677935481071472, + "learning_rate": 0.0006, + "loss": 5.290632247924805, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.05244941636919975, + "learning_rate": 0.0006, + "loss": 5.353776454925537, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.03900328651070595, + "learning_rate": 0.0006, + "loss": 5.284013748168945, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.03373100981116295, + "learning_rate": 0.0006, + "loss": 5.3313751220703125, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.03497432917356491, + "learning_rate": 0.0006, + "loss": 5.395255088806152, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.799665451049805, + "eval_runtime": 81.3525, + "eval_samples_per_second": 30.018, + "eval_steps_per_second": 0.946, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.039509695023298264, + "learning_rate": 0.0006, + "loss": 5.298876762390137, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.034599535167217255, + "learning_rate": 0.0006, + "loss": 5.264098167419434, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.029622144997119904, + "learning_rate": 0.0006, + "loss": 5.21397066116333, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.02309993840754032, + "learning_rate": 0.0006, + "loss": 5.227992057800293, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.024507572874426842, + "learning_rate": 0.0006, + "loss": 5.192324161529541, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.018214358016848564, + "learning_rate": 0.0006, + "loss": 5.23871374130249, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.01937488280236721, + "learning_rate": 0.0006, + "loss": 5.227265357971191, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.017225295305252075, + "learning_rate": 0.0006, + "loss": 5.232108116149902, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.016314376145601273, + "learning_rate": 0.0006, + "loss": 5.2288360595703125, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.014701297506690025, + "learning_rate": 0.0006, + "loss": 5.297238349914551, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.015610086731612682, + "learning_rate": 0.0006, + "loss": 5.148407936096191, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.014056755229830742, + "learning_rate": 0.0006, + "loss": 5.208695411682129, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.012802829965949059, + "learning_rate": 0.0006, + "loss": 5.229208946228027, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.011958962306380272, + "learning_rate": 0.0006, + "loss": 5.201716423034668, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.012841588817536831, + "learning_rate": 0.0006, + "loss": 5.240131378173828, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.014653137885034084, + "learning_rate": 0.0006, + "loss": 5.146340370178223, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.010446185246109962, + "learning_rate": 0.0006, + "loss": 5.21809720993042, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.01092249620705843, + "learning_rate": 0.0006, + "loss": 5.135411739349365, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.010561129078269005, + "learning_rate": 0.0006, + "loss": 5.194492340087891, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.010118645615875721, + "learning_rate": 0.0006, + "loss": 5.235012531280518, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.01081101130694151, + "learning_rate": 0.0006, + "loss": 5.162468910217285, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.010701699182391167, + "learning_rate": 0.0006, + "loss": 5.217264175415039, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.010495852679014206, + "learning_rate": 0.0006, + "loss": 5.202086925506592, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.009838009253144264, + "learning_rate": 0.0006, + "loss": 5.201089859008789, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.011280231177806854, + "learning_rate": 0.0006, + "loss": 5.1775336265563965, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.011244936846196651, + "learning_rate": 0.0006, + "loss": 5.185998439788818, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.01153804361820221, + "learning_rate": 0.0006, + "loss": 5.132739067077637, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010817309841513634, + "learning_rate": 0.0006, + "loss": 5.268172264099121, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.011870342306792736, + "learning_rate": 0.0006, + "loss": 5.154515266418457, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.011373576708137989, + "learning_rate": 0.0006, + "loss": 5.160059928894043, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.01116643100976944, + "learning_rate": 0.0006, + "loss": 5.079656600952148, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.01202121190726757, + "learning_rate": 0.0006, + "loss": 5.205666542053223, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.013683444820344448, + "learning_rate": 0.0006, + "loss": 5.168977737426758, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.012892362661659718, + "learning_rate": 0.0006, + "loss": 5.168679237365723, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.01123902853578329, + "learning_rate": 0.0006, + "loss": 5.143795013427734, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.010333497077226639, + "learning_rate": 0.0006, + "loss": 5.170592784881592, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.011278851889073849, + "learning_rate": 0.0006, + "loss": 5.162395477294922, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.013309221714735031, + "learning_rate": 0.0006, + "loss": 5.0712103843688965, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.014573273248970509, + "learning_rate": 0.0006, + "loss": 5.098300933837891, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.01755373552441597, + "learning_rate": 0.0006, + "loss": 5.081066131591797, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.02018735185265541, + "learning_rate": 0.0006, + "loss": 5.173007965087891, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.023258069530129433, + "learning_rate": 0.0006, + "loss": 5.107330322265625, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.0250930804759264, + "learning_rate": 0.0006, + "loss": 5.193347930908203, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.02333216182887554, + "learning_rate": 0.0006, + "loss": 5.160326957702637, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.019765766337513924, + "learning_rate": 0.0006, + "loss": 5.05100679397583, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.020395414903759956, + "learning_rate": 0.0006, + "loss": 5.157297611236572, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.019553566351532936, + "learning_rate": 0.0006, + "loss": 5.158632755279541, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.018838031217455864, + "learning_rate": 0.0006, + "loss": 5.167074203491211, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.020884843543171883, + "learning_rate": 0.0006, + "loss": 5.238006114959717, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.021462196484208107, + "learning_rate": 0.0006, + "loss": 5.1703901290893555, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.021178999915719032, + "learning_rate": 0.0006, + "loss": 5.109465599060059, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.022374693304300308, + "learning_rate": 0.0006, + "loss": 5.195014953613281, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.021914934739470482, + "learning_rate": 0.0006, + "loss": 5.15154504776001, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.017769858241081238, + "learning_rate": 0.0006, + "loss": 5.119076251983643, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.02018590457737446, + "learning_rate": 0.0006, + "loss": 5.087010860443115, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.023143045604228973, + "learning_rate": 0.0006, + "loss": 5.087374687194824, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.02214638516306877, + "learning_rate": 0.0006, + "loss": 5.043274879455566, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.024159753695130348, + "learning_rate": 0.0006, + "loss": 5.11695671081543, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.023281913250684738, + "learning_rate": 0.0006, + "loss": 5.159364700317383, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.02846658229827881, + "learning_rate": 0.0006, + "loss": 5.0652570724487305, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.03165476396679878, + "learning_rate": 0.0006, + "loss": 5.136387825012207, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.02582591027021408, + "learning_rate": 0.0006, + "loss": 5.059978485107422, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.025306783616542816, + "learning_rate": 0.0006, + "loss": 5.123497486114502, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.02598070539534092, + "learning_rate": 0.0006, + "loss": 5.155566215515137, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.026244031265378, + "learning_rate": 0.0006, + "loss": 5.0871381759643555, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.025219639763236046, + "learning_rate": 0.0006, + "loss": 5.102408409118652, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.023916568607091904, + "learning_rate": 0.0006, + "loss": 5.131472587585449, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.019619153812527657, + "learning_rate": 0.0006, + "loss": 5.14764404296875, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.019982660189270973, + "learning_rate": 0.0006, + "loss": 5.0941972732543945, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.018386343494057655, + "learning_rate": 0.0006, + "loss": 5.138663291931152, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.021332457661628723, + "learning_rate": 0.0006, + "loss": 5.141696453094482, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.021462492644786835, + "learning_rate": 0.0006, + "loss": 5.049698352813721, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.544618129730225, + "eval_runtime": 72.1254, + "eval_samples_per_second": 33.858, + "eval_steps_per_second": 1.068, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.02191290818154812, + "learning_rate": 0.0006, + "loss": 5.0831732749938965, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.02411566860973835, + "learning_rate": 0.0006, + "loss": 5.116334915161133, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.027553394436836243, + "learning_rate": 0.0006, + "loss": 5.050045013427734, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.02357514388859272, + "learning_rate": 0.0006, + "loss": 5.091012001037598, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.021199429407715797, + "learning_rate": 0.0006, + "loss": 5.104509353637695, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.019254926592111588, + "learning_rate": 0.0006, + "loss": 5.141340255737305, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.01811135932803154, + "learning_rate": 0.0006, + "loss": 5.046824932098389, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.017739299684762955, + "learning_rate": 0.0006, + "loss": 5.045833587646484, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.018534071743488312, + "learning_rate": 0.0006, + "loss": 4.9618401527404785, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.016690023243427277, + "learning_rate": 0.0006, + "loss": 5.165715217590332, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.018630729988217354, + "learning_rate": 0.0006, + "loss": 5.082975387573242, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.01884593442082405, + "learning_rate": 0.0006, + "loss": 5.048741817474365, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.018500948324799538, + "learning_rate": 0.0006, + "loss": 5.105315685272217, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.020263809710741043, + "learning_rate": 0.0006, + "loss": 5.087735176086426, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.019454646855592728, + "learning_rate": 0.0006, + "loss": 5.0284810066223145, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.01877632364630699, + "learning_rate": 0.0006, + "loss": 5.0442609786987305, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.020227789878845215, + "learning_rate": 0.0006, + "loss": 5.081351280212402, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.018256092444062233, + "learning_rate": 0.0006, + "loss": 5.016656875610352, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01916426233947277, + "learning_rate": 0.0006, + "loss": 5.034103870391846, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.017990969121456146, + "learning_rate": 0.0006, + "loss": 5.039344310760498, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.018346179276704788, + "learning_rate": 0.0006, + "loss": 5.017010688781738, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.01584712415933609, + "learning_rate": 0.0006, + "loss": 4.992602348327637, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.014064064249396324, + "learning_rate": 0.0006, + "loss": 5.020243167877197, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.014658658765256405, + "learning_rate": 0.0006, + "loss": 4.995723724365234, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.015613174997270107, + "learning_rate": 0.0006, + "loss": 5.0587310791015625, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.016534067690372467, + "learning_rate": 0.0006, + "loss": 4.943386077880859, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.016116850078105927, + "learning_rate": 0.0006, + "loss": 5.117921829223633, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.014806007035076618, + "learning_rate": 0.0006, + "loss": 5.077454566955566, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.013742697425186634, + "learning_rate": 0.0006, + "loss": 5.059109687805176, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.01445091050118208, + "learning_rate": 0.0006, + "loss": 4.97079610824585, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.014255733229219913, + "learning_rate": 0.0006, + "loss": 5.056376934051514, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.014356110244989395, + "learning_rate": 0.0006, + "loss": 5.095856666564941, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.015831749886274338, + "learning_rate": 0.0006, + "loss": 5.063546180725098, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.01640896685421467, + "learning_rate": 0.0006, + "loss": 5.038229942321777, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.017240721732378006, + "learning_rate": 0.0006, + "loss": 4.989873886108398, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.020683523267507553, + "learning_rate": 0.0006, + "loss": 4.945237159729004, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.02303830161690712, + "learning_rate": 0.0006, + "loss": 5.039238929748535, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.02372933179140091, + "learning_rate": 0.0006, + "loss": 5.005778789520264, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.02596980705857277, + "learning_rate": 0.0006, + "loss": 5.047043800354004, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.03045373223721981, + "learning_rate": 0.0006, + "loss": 5.07199764251709, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.033439598977565765, + "learning_rate": 0.0006, + "loss": 5.008430480957031, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.027381205931305885, + "learning_rate": 0.0006, + "loss": 5.10188102722168, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.029322044923901558, + "learning_rate": 0.0006, + "loss": 5.0966997146606445, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.02821085974574089, + "learning_rate": 0.0006, + "loss": 4.9426422119140625, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.03356366604566574, + "learning_rate": 0.0006, + "loss": 5.083784580230713, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.03138979524374008, + "learning_rate": 0.0006, + "loss": 5.063078880310059, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.02739642933011055, + "learning_rate": 0.0006, + "loss": 5.0485687255859375, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.026842772960662842, + "learning_rate": 0.0006, + "loss": 5.040291786193848, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.03270867466926575, + "learning_rate": 0.0006, + "loss": 5.118686199188232, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.027592353522777557, + "learning_rate": 0.0006, + "loss": 5.05806303024292, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.029291553422808647, + "learning_rate": 0.0006, + "loss": 5.048781394958496, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.027100825682282448, + "learning_rate": 0.0006, + "loss": 5.09072208404541, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.02333410270512104, + "learning_rate": 0.0006, + "loss": 5.0193939208984375, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.022513234987854958, + "learning_rate": 0.0006, + "loss": 4.96424674987793, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.01899099536240101, + "learning_rate": 0.0006, + "loss": 4.982175350189209, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.0192430317401886, + "learning_rate": 0.0006, + "loss": 5.094186782836914, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.02133013680577278, + "learning_rate": 0.0006, + "loss": 5.04017448425293, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.019499894231557846, + "learning_rate": 0.0006, + "loss": 5.058502197265625, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.016153644770383835, + "learning_rate": 0.0006, + "loss": 5.006505012512207, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.016342492774128914, + "learning_rate": 0.0006, + "loss": 4.998578071594238, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.014268498867750168, + "learning_rate": 0.0006, + "loss": 5.099554538726807, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.014293930493295193, + "learning_rate": 0.0006, + "loss": 4.929508686065674, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.013110974803566933, + "learning_rate": 0.0006, + "loss": 4.981067657470703, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.014049272052943707, + "learning_rate": 0.0006, + "loss": 4.97393274307251, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.011994240805506706, + "learning_rate": 0.0006, + "loss": 4.9473419189453125, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.013252398930490017, + "learning_rate": 0.0006, + "loss": 4.965734481811523, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.013359254226088524, + "learning_rate": 0.0006, + "loss": 4.968218803405762, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.01238918025046587, + "learning_rate": 0.0006, + "loss": 5.07440710067749, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.014492994174361229, + "learning_rate": 0.0006, + "loss": 5.0300493240356445, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.01550526823848486, + "learning_rate": 0.0006, + "loss": 5.031158924102783, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.016029849648475647, + "learning_rate": 0.0006, + "loss": 5.000946521759033, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.0141118373721838, + "learning_rate": 0.0006, + "loss": 5.079067230224609, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 5.463352203369141, + "eval_runtime": 79.3035, + "eval_samples_per_second": 30.793, + "eval_steps_per_second": 0.971, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.015567110851407051, + "learning_rate": 0.0006, + "loss": 5.006353855133057, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.017325397580862045, + "learning_rate": 0.0006, + "loss": 5.022473335266113, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.018784036859869957, + "learning_rate": 0.0006, + "loss": 4.922494888305664, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.020597755908966064, + "learning_rate": 0.0006, + "loss": 5.045244216918945, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.023631447926163673, + "learning_rate": 0.0006, + "loss": 4.916990280151367, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.02467048540711403, + "learning_rate": 0.0006, + "loss": 5.0631208419799805, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.024088503792881966, + "learning_rate": 0.0006, + "loss": 4.917226791381836, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.022297589108347893, + "learning_rate": 0.0006, + "loss": 4.940986156463623, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.02313072420656681, + "learning_rate": 0.0006, + "loss": 5.013306617736816, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.028538871556520462, + "learning_rate": 0.0006, + "loss": 4.99459171295166, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.029122058302164078, + "learning_rate": 0.0006, + "loss": 5.000114440917969, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.022197917103767395, + "learning_rate": 0.0006, + "loss": 4.945086479187012, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.021084975451231003, + "learning_rate": 0.0006, + "loss": 4.920244216918945, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.022474950179457664, + "learning_rate": 0.0006, + "loss": 5.005110740661621, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.02403559722006321, + "learning_rate": 0.0006, + "loss": 5.04328727722168, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.026955733075737953, + "learning_rate": 0.0006, + "loss": 4.962291717529297, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.024654097855091095, + "learning_rate": 0.0006, + "loss": 5.0805840492248535, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.01992264948785305, + "learning_rate": 0.0006, + "loss": 5.044548034667969, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01696479320526123, + "learning_rate": 0.0006, + "loss": 4.912850856781006, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.017840469256043434, + "learning_rate": 0.0006, + "loss": 5.0556793212890625, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.01623854599893093, + "learning_rate": 0.0006, + "loss": 5.040976524353027, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.016552211716771126, + "learning_rate": 0.0006, + "loss": 4.971366882324219, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.015316464938223362, + "learning_rate": 0.0006, + "loss": 4.952664375305176, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.01712590828537941, + "learning_rate": 0.0006, + "loss": 4.969407081604004, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.015214908868074417, + "learning_rate": 0.0006, + "loss": 4.9463911056518555, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.014866896905004978, + "learning_rate": 0.0006, + "loss": 5.071913242340088, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.01533361803740263, + "learning_rate": 0.0006, + "loss": 5.0071916580200195, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.014022483490407467, + "learning_rate": 0.0006, + "loss": 4.954067707061768, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.015019413083791733, + "learning_rate": 0.0006, + "loss": 4.9905900955200195, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.015474580228328705, + "learning_rate": 0.0006, + "loss": 4.999631881713867, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.020479638129472733, + "learning_rate": 0.0006, + "loss": 5.0200700759887695, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.0237763449549675, + "learning_rate": 0.0006, + "loss": 5.008729457855225, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.022637002170085907, + "learning_rate": 0.0006, + "loss": 4.940751075744629, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.019718417897820473, + "learning_rate": 0.0006, + "loss": 4.9788737297058105, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.023200126364827156, + "learning_rate": 0.0006, + "loss": 4.906790256500244, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.025976872071623802, + "learning_rate": 0.0006, + "loss": 5.002409934997559, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.023092452436685562, + "learning_rate": 0.0006, + "loss": 4.90887451171875, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.01955891214311123, + "learning_rate": 0.0006, + "loss": 4.995199203491211, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.02019096165895462, + "learning_rate": 0.0006, + "loss": 4.908695697784424, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.019376836717128754, + "learning_rate": 0.0006, + "loss": 4.872652053833008, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.019068755209445953, + "learning_rate": 0.0006, + "loss": 4.952585220336914, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.016856160014867783, + "learning_rate": 0.0006, + "loss": 4.927958965301514, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.017544621601700783, + "learning_rate": 0.0006, + "loss": 4.922074794769287, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.01814766600728035, + "learning_rate": 0.0006, + "loss": 4.884115219116211, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.01756724901497364, + "learning_rate": 0.0006, + "loss": 4.969221115112305, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.016586478799581528, + "learning_rate": 0.0006, + "loss": 4.897098541259766, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.0165987778455019, + "learning_rate": 0.0006, + "loss": 4.8751044273376465, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.019033856689929962, + "learning_rate": 0.0006, + "loss": 4.894170761108398, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.016953010112047195, + "learning_rate": 0.0006, + "loss": 4.918987274169922, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016293920576572418, + "learning_rate": 0.0006, + "loss": 4.949659824371338, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.019263453781604767, + "learning_rate": 0.0006, + "loss": 4.923409461975098, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.01789483241736889, + "learning_rate": 0.0006, + "loss": 4.894615173339844, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01610061153769493, + "learning_rate": 0.0006, + "loss": 4.9311604499816895, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01674770377576351, + "learning_rate": 0.0006, + "loss": 4.938290596008301, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.01579914800822735, + "learning_rate": 0.0006, + "loss": 4.936288833618164, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.015871549025177956, + "learning_rate": 0.0006, + "loss": 4.970236301422119, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.014079514890909195, + "learning_rate": 0.0006, + "loss": 4.935883045196533, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.015348718501627445, + "learning_rate": 0.0006, + "loss": 4.919315814971924, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.0189143568277359, + "learning_rate": 0.0006, + "loss": 4.972513198852539, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.021286798641085625, + "learning_rate": 0.0006, + "loss": 4.89121675491333, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.0212328489869833, + "learning_rate": 0.0006, + "loss": 4.895394802093506, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.020312150940299034, + "learning_rate": 0.0006, + "loss": 4.943729400634766, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.02121325396001339, + "learning_rate": 0.0006, + "loss": 4.9322638511657715, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.02030770666897297, + "learning_rate": 0.0006, + "loss": 4.8969550132751465, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.02408582717180252, + "learning_rate": 0.0006, + "loss": 4.94023323059082, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.025230303406715393, + "learning_rate": 0.0006, + "loss": 4.964077472686768, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.025903481990098953, + "learning_rate": 0.0006, + "loss": 4.9319024085998535, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.02517629973590374, + "learning_rate": 0.0006, + "loss": 4.9937262535095215, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.023874718695878983, + "learning_rate": 0.0006, + "loss": 4.916127681732178, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.02531718835234642, + "learning_rate": 0.0006, + "loss": 4.949875831604004, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.026585623621940613, + "learning_rate": 0.0006, + "loss": 4.883786201477051, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.029191339388489723, + "learning_rate": 0.0006, + "loss": 4.955547332763672, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 5.376297473907471, + "eval_runtime": 57.4412, + "eval_samples_per_second": 42.513, + "eval_steps_per_second": 1.341, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.020493783056735992, + "learning_rate": 0.0006, + "loss": 4.887005805969238, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.022943608462810516, + "learning_rate": 0.0006, + "loss": 4.9878644943237305, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.026248518377542496, + "learning_rate": 0.0006, + "loss": 4.929098129272461, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.027000896632671356, + "learning_rate": 0.0006, + "loss": 4.97839879989624, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.026088254526257515, + "learning_rate": 0.0006, + "loss": 5.037898540496826, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.027397966012358665, + "learning_rate": 0.0006, + "loss": 4.8592939376831055, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.024916915223002434, + "learning_rate": 0.0006, + "loss": 4.908574104309082, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.026325512677431107, + "learning_rate": 0.0006, + "loss": 4.8777313232421875, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.02321736514568329, + "learning_rate": 0.0006, + "loss": 4.950716972351074, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.020332302898168564, + "learning_rate": 0.0006, + "loss": 4.934875965118408, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.02042531967163086, + "learning_rate": 0.0006, + "loss": 4.937285423278809, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.018892595544457436, + "learning_rate": 0.0006, + "loss": 4.909456253051758, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.018573295325040817, + "learning_rate": 0.0006, + "loss": 4.886369228363037, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.019991718232631683, + "learning_rate": 0.0006, + "loss": 4.923654556274414, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.015044108964502811, + "learning_rate": 0.0006, + "loss": 4.983735084533691, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.016339287161827087, + "learning_rate": 0.0006, + "loss": 4.840668678283691, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.015570246614515781, + "learning_rate": 0.0006, + "loss": 4.88963508605957, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.014452921226620674, + "learning_rate": 0.0006, + "loss": 4.931870460510254, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.014817588031291962, + "learning_rate": 0.0006, + "loss": 4.893659591674805, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013385407626628876, + "learning_rate": 0.0006, + "loss": 4.899247646331787, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.01410810649394989, + "learning_rate": 0.0006, + "loss": 4.9700541496276855, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.015247197821736336, + "learning_rate": 0.0006, + "loss": 4.945037841796875, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.015484587289392948, + "learning_rate": 0.0006, + "loss": 4.924195289611816, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.014522048644721508, + "learning_rate": 0.0006, + "loss": 4.89838981628418, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.015260211192071438, + "learning_rate": 0.0006, + "loss": 4.845786094665527, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.015550734475255013, + "learning_rate": 0.0006, + "loss": 4.893673896789551, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.013600634410977364, + "learning_rate": 0.0006, + "loss": 4.906546592712402, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.013318133540451527, + "learning_rate": 0.0006, + "loss": 4.897005081176758, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.013978407718241215, + "learning_rate": 0.0006, + "loss": 4.850296974182129, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.014549889601767063, + "learning_rate": 0.0006, + "loss": 4.874753952026367, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.018115175887942314, + "learning_rate": 0.0006, + "loss": 4.818711280822754, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.018728114664554596, + "learning_rate": 0.0006, + "loss": 4.882294654846191, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.018031353130936623, + "learning_rate": 0.0006, + "loss": 4.828677177429199, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.017293527722358704, + "learning_rate": 0.0006, + "loss": 4.872529983520508, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.015710802748799324, + "learning_rate": 0.0006, + "loss": 4.921950817108154, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.017632456496357918, + "learning_rate": 0.0006, + "loss": 4.907940864562988, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018707087263464928, + "learning_rate": 0.0006, + "loss": 4.797819137573242, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.019626259803771973, + "learning_rate": 0.0006, + "loss": 4.795708656311035, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.018569709733128548, + "learning_rate": 0.0006, + "loss": 4.876678466796875, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.01720404624938965, + "learning_rate": 0.0006, + "loss": 4.922236442565918, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.018940523266792297, + "learning_rate": 0.0006, + "loss": 4.892475605010986, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.020524146035313606, + "learning_rate": 0.0006, + "loss": 4.814975738525391, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.017521994188427925, + "learning_rate": 0.0006, + "loss": 4.912031173706055, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.019237685948610306, + "learning_rate": 0.0006, + "loss": 4.8619184494018555, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.018941357731819153, + "learning_rate": 0.0006, + "loss": 4.983427047729492, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.01969471573829651, + "learning_rate": 0.0006, + "loss": 4.877185344696045, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.01886482909321785, + "learning_rate": 0.0006, + "loss": 4.858532905578613, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.018075549975037575, + "learning_rate": 0.0006, + "loss": 4.85374641418457, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.018726196140050888, + "learning_rate": 0.0006, + "loss": 4.9164719581604, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.020012663677334785, + "learning_rate": 0.0006, + "loss": 4.790283203125, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.01965046115219593, + "learning_rate": 0.0006, + "loss": 4.833796501159668, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.020102867856621742, + "learning_rate": 0.0006, + "loss": 4.835939407348633, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.01858644187450409, + "learning_rate": 0.0006, + "loss": 4.819847106933594, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.02036326937377453, + "learning_rate": 0.0006, + "loss": 4.9040985107421875, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.023322949185967445, + "learning_rate": 0.0006, + "loss": 4.824136734008789, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.023923879489302635, + "learning_rate": 0.0006, + "loss": 4.91016960144043, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.02464689128100872, + "learning_rate": 0.0006, + "loss": 4.949073791503906, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.02039799466729164, + "learning_rate": 0.0006, + "loss": 4.93691349029541, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.019222574308514595, + "learning_rate": 0.0006, + "loss": 4.936672210693359, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.017282437533140182, + "learning_rate": 0.0006, + "loss": 4.818748950958252, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.021188177168369293, + "learning_rate": 0.0006, + "loss": 4.880117416381836, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.027654651552438736, + "learning_rate": 0.0006, + "loss": 4.827604293823242, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.028346989303827286, + "learning_rate": 0.0006, + "loss": 4.872309684753418, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.02339531108736992, + "learning_rate": 0.0006, + "loss": 4.87053918838501, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.02051672339439392, + "learning_rate": 0.0006, + "loss": 4.901670455932617, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.02038448490202427, + "learning_rate": 0.0006, + "loss": 4.822744846343994, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.0207635797560215, + "learning_rate": 0.0006, + "loss": 4.83148717880249, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.020679432898759842, + "learning_rate": 0.0006, + "loss": 4.870957374572754, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.019351216033101082, + "learning_rate": 0.0006, + "loss": 4.956622123718262, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.018666930496692657, + "learning_rate": 0.0006, + "loss": 4.853818893432617, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.019687574356794357, + "learning_rate": 0.0006, + "loss": 4.827330589294434, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.020736118778586388, + "learning_rate": 0.0006, + "loss": 4.856723785400391, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 5.262957572937012, + "eval_runtime": 59.9518, + "eval_samples_per_second": 40.733, + "eval_steps_per_second": 1.284, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.019920889288187027, + "learning_rate": 0.0006, + "loss": 4.925846099853516, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.020640799775719643, + "learning_rate": 0.0006, + "loss": 4.83962345123291, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.019742414355278015, + "learning_rate": 0.0006, + "loss": 4.935937881469727, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.018709300085902214, + "learning_rate": 0.0006, + "loss": 4.866799831390381, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.01504162885248661, + "learning_rate": 0.0006, + "loss": 4.810856819152832, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.014874175190925598, + "learning_rate": 0.0006, + "loss": 4.881101608276367, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.017353734001517296, + "learning_rate": 0.0006, + "loss": 4.846688747406006, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.017002031207084656, + "learning_rate": 0.0006, + "loss": 4.807414531707764, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.016849525272846222, + "learning_rate": 0.0006, + "loss": 4.811706066131592, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.017146047204732895, + "learning_rate": 0.0006, + "loss": 4.811017990112305, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.018497949466109276, + "learning_rate": 0.0006, + "loss": 4.8794403076171875, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.02204900048673153, + "learning_rate": 0.0006, + "loss": 4.917919635772705, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.02386346273124218, + "learning_rate": 0.0006, + "loss": 4.941521644592285, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.019803019240498543, + "learning_rate": 0.0006, + "loss": 4.772615432739258, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.01547364704310894, + "learning_rate": 0.0006, + "loss": 4.817747592926025, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.015355406329035759, + "learning_rate": 0.0006, + "loss": 4.9097771644592285, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.013763652183115482, + "learning_rate": 0.0006, + "loss": 4.790132522583008, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.015364146791398525, + "learning_rate": 0.0006, + "loss": 4.876467704772949, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01718062348663807, + "learning_rate": 0.0006, + "loss": 4.871622562408447, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.01469398569315672, + "learning_rate": 0.0006, + "loss": 4.829773426055908, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.015336665324866772, + "learning_rate": 0.0006, + "loss": 4.886260032653809, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.01615484617650509, + "learning_rate": 0.0006, + "loss": 4.840301990509033, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.017155757173895836, + "learning_rate": 0.0006, + "loss": 4.7739787101745605, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.016230205073952675, + "learning_rate": 0.0006, + "loss": 4.790524005889893, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.016484249383211136, + "learning_rate": 0.0006, + "loss": 4.948139190673828, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.016610991209745407, + "learning_rate": 0.0006, + "loss": 4.872422218322754, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.017981581389904022, + "learning_rate": 0.0006, + "loss": 4.873099327087402, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.020764052867889404, + "learning_rate": 0.0006, + "loss": 4.840051174163818, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.02078232914209366, + "learning_rate": 0.0006, + "loss": 4.776272773742676, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.020214475691318512, + "learning_rate": 0.0006, + "loss": 4.83798360824585, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.02398747019469738, + "learning_rate": 0.0006, + "loss": 4.770004749298096, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.024379247799515724, + "learning_rate": 0.0006, + "loss": 4.905134201049805, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.020821094512939453, + "learning_rate": 0.0006, + "loss": 4.821808815002441, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.018303627148270607, + "learning_rate": 0.0006, + "loss": 4.7170562744140625, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.01970406249165535, + "learning_rate": 0.0006, + "loss": 4.872300148010254, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.01975054107606411, + "learning_rate": 0.0006, + "loss": 4.850522994995117, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.018307320773601532, + "learning_rate": 0.0006, + "loss": 4.80927038192749, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.0156140411272645, + "learning_rate": 0.0006, + "loss": 4.816516399383545, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.0153433782979846, + "learning_rate": 0.0006, + "loss": 4.719747066497803, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.014209619723260403, + "learning_rate": 0.0006, + "loss": 4.838152885437012, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.015228115022182465, + "learning_rate": 0.0006, + "loss": 4.744102954864502, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.015567339025437832, + "learning_rate": 0.0006, + "loss": 4.789331436157227, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.01851458102464676, + "learning_rate": 0.0006, + "loss": 4.847858905792236, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.01986544393002987, + "learning_rate": 0.0006, + "loss": 4.826117992401123, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.0173768550157547, + "learning_rate": 0.0006, + "loss": 4.838115215301514, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.01780487783253193, + "learning_rate": 0.0006, + "loss": 4.820503234863281, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.018898125737905502, + "learning_rate": 0.0006, + "loss": 4.8488359451293945, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.016740918159484863, + "learning_rate": 0.0006, + "loss": 4.815840244293213, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.017316583544015884, + "learning_rate": 0.0006, + "loss": 4.787709712982178, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.01631874032318592, + "learning_rate": 0.0006, + "loss": 4.807087421417236, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.018026666715741158, + "learning_rate": 0.0006, + "loss": 4.810066223144531, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01884373649954796, + "learning_rate": 0.0006, + "loss": 4.852358818054199, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.01915278099477291, + "learning_rate": 0.0006, + "loss": 4.850739002227783, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.01981119066476822, + "learning_rate": 0.0006, + "loss": 4.794268608093262, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.020773326978087425, + "learning_rate": 0.0006, + "loss": 4.789695739746094, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.0198501106351614, + "learning_rate": 0.0006, + "loss": 4.812196731567383, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.016126036643981934, + "learning_rate": 0.0006, + "loss": 4.8933563232421875, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.015466060489416122, + "learning_rate": 0.0006, + "loss": 4.765222549438477, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.01579921506345272, + "learning_rate": 0.0006, + "loss": 4.859963893890381, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.015828445553779602, + "learning_rate": 0.0006, + "loss": 4.8465094566345215, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.01661691628396511, + "learning_rate": 0.0006, + "loss": 4.766627311706543, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01876002363860607, + "learning_rate": 0.0006, + "loss": 4.792072296142578, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.01939656026661396, + "learning_rate": 0.0006, + "loss": 4.8140668869018555, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.019839171320199966, + "learning_rate": 0.0006, + "loss": 4.909534454345703, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.020329177379608154, + "learning_rate": 0.0006, + "loss": 4.803993225097656, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.019034743309020996, + "learning_rate": 0.0006, + "loss": 4.822991847991943, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.018672214820981026, + "learning_rate": 0.0006, + "loss": 4.771214485168457, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.019761107861995697, + "learning_rate": 0.0006, + "loss": 4.84083366394043, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.018283944576978683, + "learning_rate": 0.0006, + "loss": 4.782181739807129, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.02122645638883114, + "learning_rate": 0.0006, + "loss": 4.7942962646484375, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.02146247588098049, + "learning_rate": 0.0006, + "loss": 4.7972211837768555, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.022740444168448448, + "learning_rate": 0.0006, + "loss": 4.778811931610107, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 5.211768627166748, + "eval_runtime": 58.8095, + "eval_samples_per_second": 41.524, + "eval_steps_per_second": 1.309, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.02356165461242199, + "learning_rate": 0.0006, + "loss": 4.815979480743408, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.024365399032831192, + "learning_rate": 0.0006, + "loss": 4.731906890869141, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.02496817521750927, + "learning_rate": 0.0006, + "loss": 4.798555850982666, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.023278620094060898, + "learning_rate": 0.0006, + "loss": 4.764307975769043, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.020955104380846024, + "learning_rate": 0.0006, + "loss": 4.766234874725342, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.021213296800851822, + "learning_rate": 0.0006, + "loss": 4.856618881225586, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.022735338658094406, + "learning_rate": 0.0006, + "loss": 4.853489875793457, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.024209095165133476, + "learning_rate": 0.0006, + "loss": 4.858719825744629, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.025767242535948753, + "learning_rate": 0.0006, + "loss": 4.871748447418213, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.021460559219121933, + "learning_rate": 0.0006, + "loss": 4.686347961425781, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.019937612116336823, + "learning_rate": 0.0006, + "loss": 4.792636871337891, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.022846097126603127, + "learning_rate": 0.0006, + "loss": 4.818110466003418, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.020246420055627823, + "learning_rate": 0.0006, + "loss": 4.855318546295166, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.017798585817217827, + "learning_rate": 0.0006, + "loss": 4.776525497436523, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.01727782003581524, + "learning_rate": 0.0006, + "loss": 4.850035667419434, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01917205937206745, + "learning_rate": 0.0006, + "loss": 4.787570953369141, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.020757799968123436, + "learning_rate": 0.0006, + "loss": 4.797884941101074, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.019872257485985756, + "learning_rate": 0.0006, + "loss": 4.823526382446289, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.017858969047665596, + "learning_rate": 0.0006, + "loss": 4.774180889129639, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.01774270087480545, + "learning_rate": 0.0006, + "loss": 4.739023208618164, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.018475241959095, + "learning_rate": 0.0006, + "loss": 4.8812456130981445, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.018330633640289307, + "learning_rate": 0.0006, + "loss": 4.738748073577881, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.016155779361724854, + "learning_rate": 0.0006, + "loss": 4.81126594543457, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.016643118113279343, + "learning_rate": 0.0006, + "loss": 4.796675682067871, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.0177546925842762, + "learning_rate": 0.0006, + "loss": 4.7574028968811035, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.017025742679834366, + "learning_rate": 0.0006, + "loss": 4.737842559814453, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.01645711250603199, + "learning_rate": 0.0006, + "loss": 4.762307643890381, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.015202434733510017, + "learning_rate": 0.0006, + "loss": 4.844311237335205, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.013969679363071918, + "learning_rate": 0.0006, + "loss": 4.757219314575195, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.013256058096885681, + "learning_rate": 0.0006, + "loss": 4.748269081115723, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.014061232097446918, + "learning_rate": 0.0006, + "loss": 4.796285152435303, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.014178436249494553, + "learning_rate": 0.0006, + "loss": 4.773000717163086, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.015295447781682014, + "learning_rate": 0.0006, + "loss": 4.785943031311035, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.014845133759081364, + "learning_rate": 0.0006, + "loss": 4.831053733825684, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.015798132866621017, + "learning_rate": 0.0006, + "loss": 4.87584114074707, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.017797362059354782, + "learning_rate": 0.0006, + "loss": 4.769850254058838, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.01941198669373989, + "learning_rate": 0.0006, + "loss": 4.71127986907959, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.01806589961051941, + "learning_rate": 0.0006, + "loss": 4.718900203704834, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.01684599742293358, + "learning_rate": 0.0006, + "loss": 4.77104377746582, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.017405332997441292, + "learning_rate": 0.0006, + "loss": 4.751084327697754, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.017482969909906387, + "learning_rate": 0.0006, + "loss": 4.771002769470215, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.018441764637827873, + "learning_rate": 0.0006, + "loss": 4.77370548248291, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.019556425511837006, + "learning_rate": 0.0006, + "loss": 4.791923999786377, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.022014815360307693, + "learning_rate": 0.0006, + "loss": 4.7738189697265625, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.0217057466506958, + "learning_rate": 0.0006, + "loss": 4.738826274871826, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01779630407691002, + "learning_rate": 0.0006, + "loss": 4.771679401397705, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.017392629757523537, + "learning_rate": 0.0006, + "loss": 4.796988487243652, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.016696074977517128, + "learning_rate": 0.0006, + "loss": 4.720436096191406, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.01689007878303528, + "learning_rate": 0.0006, + "loss": 4.798285961151123, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.018068408593535423, + "learning_rate": 0.0006, + "loss": 4.764993190765381, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.01882725954055786, + "learning_rate": 0.0006, + "loss": 4.751399040222168, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.01765439659357071, + "learning_rate": 0.0006, + "loss": 4.8288493156433105, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.018233707174658775, + "learning_rate": 0.0006, + "loss": 4.762392997741699, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.019315950572490692, + "learning_rate": 0.0006, + "loss": 4.775703430175781, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.01821810007095337, + "learning_rate": 0.0006, + "loss": 4.8006181716918945, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.018199540674686432, + "learning_rate": 0.0006, + "loss": 4.774545192718506, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.019126253202557564, + "learning_rate": 0.0006, + "loss": 4.796529769897461, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.019279757514595985, + "learning_rate": 0.0006, + "loss": 4.7565107345581055, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.017493106424808502, + "learning_rate": 0.0006, + "loss": 4.74436616897583, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.01618535816669464, + "learning_rate": 0.0006, + "loss": 4.8202362060546875, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.014561207965016365, + "learning_rate": 0.0006, + "loss": 4.8702802658081055, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.014859725721180439, + "learning_rate": 0.0006, + "loss": 4.740710735321045, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014770046807825565, + "learning_rate": 0.0006, + "loss": 4.726039409637451, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.01728186383843422, + "learning_rate": 0.0006, + "loss": 4.751114368438721, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.016778577119112015, + "learning_rate": 0.0006, + "loss": 4.841277122497559, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.017361296340823174, + "learning_rate": 0.0006, + "loss": 4.753442764282227, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.020100396126508713, + "learning_rate": 0.0006, + "loss": 4.657037734985352, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.024198055267333984, + "learning_rate": 0.0006, + "loss": 4.868254661560059, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.024044830352067947, + "learning_rate": 0.0006, + "loss": 4.760613918304443, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.02042238786816597, + "learning_rate": 0.0006, + "loss": 4.729440212249756, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.02100338228046894, + "learning_rate": 0.0006, + "loss": 4.751463413238525, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.020707951858639717, + "learning_rate": 0.0006, + "loss": 4.8115129470825195, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 5.2392497062683105, + "eval_runtime": 61.5697, + "eval_samples_per_second": 39.662, + "eval_steps_per_second": 1.251, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.020968902856111526, + "learning_rate": 0.0006, + "loss": 4.749300479888916, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.019791511818766594, + "learning_rate": 0.0006, + "loss": 4.752042770385742, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.01768038608133793, + "learning_rate": 0.0006, + "loss": 4.804799556732178, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.015827463939785957, + "learning_rate": 0.0006, + "loss": 4.746969699859619, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.01784376986324787, + "learning_rate": 0.0006, + "loss": 4.709347724914551, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.017368916422128677, + "learning_rate": 0.0006, + "loss": 4.701427459716797, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.017079368233680725, + "learning_rate": 0.0006, + "loss": 4.740375518798828, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.01675686053931713, + "learning_rate": 0.0006, + "loss": 4.7559404373168945, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.016627594828605652, + "learning_rate": 0.0006, + "loss": 4.686899185180664, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.01687782071530819, + "learning_rate": 0.0006, + "loss": 4.694512367248535, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.017982905730605125, + "learning_rate": 0.0006, + "loss": 4.771855354309082, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.018535513430833817, + "learning_rate": 0.0006, + "loss": 4.67527437210083, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.018792208284139633, + "learning_rate": 0.0006, + "loss": 4.776930332183838, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.019609590992331505, + "learning_rate": 0.0006, + "loss": 4.740622520446777, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.020893512293696404, + "learning_rate": 0.0006, + "loss": 4.761295795440674, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.018166005611419678, + "learning_rate": 0.0006, + "loss": 4.760861396789551, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.016187280416488647, + "learning_rate": 0.0006, + "loss": 4.709270477294922, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.014516904018819332, + "learning_rate": 0.0006, + "loss": 4.706027984619141, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.014672615565359592, + "learning_rate": 0.0006, + "loss": 4.764942169189453, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.015825409442186356, + "learning_rate": 0.0006, + "loss": 4.862574100494385, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.01423126831650734, + "learning_rate": 0.0006, + "loss": 4.721527576446533, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.015230957418680191, + "learning_rate": 0.0006, + "loss": 4.789156913757324, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.014533719047904015, + "learning_rate": 0.0006, + "loss": 4.727848052978516, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.015545960515737534, + "learning_rate": 0.0006, + "loss": 4.67042350769043, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.015026269480586052, + "learning_rate": 0.0006, + "loss": 4.8094048500061035, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.016231181100010872, + "learning_rate": 0.0006, + "loss": 4.743175983428955, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.0172134917229414, + "learning_rate": 0.0006, + "loss": 4.633797645568848, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.016263330355286598, + "learning_rate": 0.0006, + "loss": 4.642855644226074, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.014997457154095173, + "learning_rate": 0.0006, + "loss": 4.6688408851623535, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015129785053431988, + "learning_rate": 0.0006, + "loss": 4.673038482666016, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.017703495919704437, + "learning_rate": 0.0006, + "loss": 4.699586868286133, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.017693769186735153, + "learning_rate": 0.0006, + "loss": 4.741053104400635, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.018453532829880714, + "learning_rate": 0.0006, + "loss": 4.696603775024414, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.018508777022361755, + "learning_rate": 0.0006, + "loss": 4.652768135070801, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016347555443644524, + "learning_rate": 0.0006, + "loss": 4.696932792663574, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.016488511115312576, + "learning_rate": 0.0006, + "loss": 4.668985843658447, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.016894331201910973, + "learning_rate": 0.0006, + "loss": 4.712052345275879, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.01606992818415165, + "learning_rate": 0.0006, + "loss": 4.809332847595215, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.01614546775817871, + "learning_rate": 0.0006, + "loss": 4.718560218811035, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.015790050849318504, + "learning_rate": 0.0006, + "loss": 4.7086992263793945, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.015570463612675667, + "learning_rate": 0.0006, + "loss": 4.701510906219482, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.015313868410885334, + "learning_rate": 0.0006, + "loss": 4.6717400550842285, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.016266439110040665, + "learning_rate": 0.0006, + "loss": 4.708248138427734, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.01676693931221962, + "learning_rate": 0.0006, + "loss": 4.708803176879883, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.016014447435736656, + "learning_rate": 0.0006, + "loss": 4.7899603843688965, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01589064858853817, + "learning_rate": 0.0006, + "loss": 4.683867454528809, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.01659962348639965, + "learning_rate": 0.0006, + "loss": 4.789435863494873, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.01895272172987461, + "learning_rate": 0.0006, + "loss": 4.753860950469971, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.019914697855710983, + "learning_rate": 0.0006, + "loss": 4.701515197753906, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.021606286987662315, + "learning_rate": 0.0006, + "loss": 4.754340171813965, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.02217685803771019, + "learning_rate": 0.0006, + "loss": 4.73709774017334, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.020211679860949516, + "learning_rate": 0.0006, + "loss": 4.6904473304748535, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.020696299150586128, + "learning_rate": 0.0006, + "loss": 4.718563079833984, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.02091088518500328, + "learning_rate": 0.0006, + "loss": 4.803199291229248, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.020736388862133026, + "learning_rate": 0.0006, + "loss": 4.735616207122803, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.018370863050222397, + "learning_rate": 0.0006, + "loss": 4.6523332595825195, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.01750843971967697, + "learning_rate": 0.0006, + "loss": 4.715566635131836, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.018078286200761795, + "learning_rate": 0.0006, + "loss": 4.734353065490723, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.018023716285824776, + "learning_rate": 0.0006, + "loss": 4.652746200561523, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.018733404576778412, + "learning_rate": 0.0006, + "loss": 4.7124176025390625, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.018829114735126495, + "learning_rate": 0.0006, + "loss": 4.720004081726074, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.019517114385962486, + "learning_rate": 0.0006, + "loss": 4.7428741455078125, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.01994675025343895, + "learning_rate": 0.0006, + "loss": 4.717165946960449, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.01924729160964489, + "learning_rate": 0.0006, + "loss": 4.749120712280273, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.018079813569784164, + "learning_rate": 0.0006, + "loss": 4.794156551361084, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.017745722085237503, + "learning_rate": 0.0006, + "loss": 4.7152910232543945, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.020233964547514915, + "learning_rate": 0.0006, + "loss": 4.7771477699279785, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.020702997222542763, + "learning_rate": 0.0006, + "loss": 4.7343339920043945, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.021028192713856697, + "learning_rate": 0.0006, + "loss": 4.7011847496032715, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.018680263310670853, + "learning_rate": 0.0006, + "loss": 4.796342849731445, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.017404882237315178, + "learning_rate": 0.0006, + "loss": 4.678404808044434, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.017616212368011475, + "learning_rate": 0.0006, + "loss": 4.792126655578613, + "step": 1440 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6.139226722664448e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-1440/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef +size 4792 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/config.json new file mode 100644 index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151674 +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1aed3be13b8db54755855a33bd4e4da31a650c8b --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecb01abdbbdb51cecfa2cf09d771518c1906251227edb20fdbfe400a9ddf08fe +size 583362376 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5aebb2357fb3cc1180e6bd78ef51e9c2157663ec --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:945d4f54df19bbbb716cfcb2dd5351ce119b256582670b18b3f3bf825f213d2b +size 1166837626 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..81d2bd9decff40322863d3c4253a6f90b727f3ce --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f123a34234139b64f0d7a1a8066d6c6e69ecc4d10c733547345f9db86126081c +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..51368e431549a274df92d5a7d67ffef092809d98 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1e2b3f9a2385ece5e2513f9f80c9c776fe9b56f608574d5f628c540850f3d4d +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..77d0825438ee7f93a226a8383ff7b82b66e7d59a --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:775f48d24f2da0edd9623b436cc53a297a5f382317102f871a0af9b07b45a983 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..6b101452eb1e045f4259258fdac0f1336b38a5f5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b255c26ded3d367afe0151ee7ddf729f4b29ddc8696bcfb44058cf024acb072b +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..e7f2e4bf5abc860ae45f2d5dbda8115542dd3611 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d1eab1d6f570939dc1e4b9d10383b7354ef244100fb2b34ff0ac8b58642b63c +size 1064 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14 +size 11424648 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/tokenizer_config.json @@ -0,0 +1,21 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>", + "<|mask|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..c539125b4bac0682d1fbedc1a03928bdb3624066 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/trainer_state.json @@ -0,0 +1,2074 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.0, + "eval_steps": 500, + "global_step": 288, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.5768774151802063, + "learning_rate": 0.0, + "loss": 12.050328254699707, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.5784570574760437, + "learning_rate": 5.999999999999999e-06, + "loss": 12.049251556396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.5607187151908875, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.921792030334473, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.46466919779777527, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.686628341674805, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.38012945652008057, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.432823181152344, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.33215776085853577, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.207542419433594, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.2928776741027832, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.000097274780273, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.26018962264060974, + "learning_rate": 4.2e-05, + "loss": 10.825748443603516, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.23717083036899567, + "learning_rate": 4.7999999999999994e-05, + "loss": 10.677122116088867, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.21641355752944946, + "learning_rate": 5.399999999999999e-05, + "loss": 10.569148063659668, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.1982956826686859, + "learning_rate": 5.9999999999999995e-05, + "loss": 10.49653434753418, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.18819282948970795, + "learning_rate": 6.599999999999999e-05, + "loss": 10.432951927185059, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.18322618305683136, + "learning_rate": 7.199999999999999e-05, + "loss": 10.387115478515625, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.18399226665496826, + "learning_rate": 7.8e-05, + "loss": 10.33381462097168, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.1833190619945526, + "learning_rate": 8.4e-05, + "loss": 10.301689147949219, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.18490834534168243, + "learning_rate": 8.999999999999999e-05, + "loss": 10.240968704223633, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.18358023464679718, + "learning_rate": 9.599999999999999e-05, + "loss": 10.194717407226562, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.1801682710647583, + "learning_rate": 0.000102, + "loss": 10.14897632598877, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.17899440228939056, + "learning_rate": 0.00010799999999999998, + "loss": 10.078763961791992, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.17666497826576233, + "learning_rate": 0.00011399999999999999, + "loss": 10.01095962524414, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.17196154594421387, + "learning_rate": 0.00011999999999999999, + "loss": 9.951203346252441, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.169203519821167, + "learning_rate": 0.00012599999999999997, + "loss": 9.871726036071777, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.16500306129455566, + "learning_rate": 0.00013199999999999998, + "loss": 9.803646087646484, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.16128124296665192, + "learning_rate": 0.000138, + "loss": 9.728487014770508, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.15904058516025543, + "learning_rate": 0.00014399999999999998, + "loss": 9.642473220825195, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.15560561418533325, + "learning_rate": 0.00015, + "loss": 9.56414794921875, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.15278612077236176, + "learning_rate": 0.000156, + "loss": 9.484230041503906, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.15263237059116364, + "learning_rate": 0.000162, + "loss": 9.383644104003906, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.14945024251937866, + "learning_rate": 0.000168, + "loss": 9.29840087890625, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.14614646136760712, + "learning_rate": 0.00017399999999999997, + "loss": 9.208931922912598, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.14295098185539246, + "learning_rate": 0.00017999999999999998, + "loss": 9.122323989868164, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.13868345320224762, + "learning_rate": 0.000186, + "loss": 9.03034496307373, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.135064035654068, + "learning_rate": 0.00019199999999999998, + "loss": 8.92818832397461, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.13097241520881653, + "learning_rate": 0.000198, + "loss": 8.834627151489258, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.12478776276111603, + "learning_rate": 0.000204, + "loss": 8.755172729492188, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.1214982345700264, + "learning_rate": 0.00020999999999999998, + "loss": 8.647891998291016, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.11958328634500504, + "learning_rate": 0.00021599999999999996, + "loss": 8.543274879455566, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.11680915951728821, + "learning_rate": 0.00022199999999999998, + "loss": 8.44540023803711, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.11491474509239197, + "learning_rate": 0.00022799999999999999, + "loss": 8.349921226501465, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.11160755902528763, + "learning_rate": 0.000234, + "loss": 8.26960563659668, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.10840637236833572, + "learning_rate": 0.00023999999999999998, + "loss": 8.192136764526367, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.10585298389196396, + "learning_rate": 0.00024599999999999996, + "loss": 8.09546947479248, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.1032542735338211, + "learning_rate": 0.00025199999999999995, + "loss": 7.991090774536133, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.0986943170428276, + "learning_rate": 0.000258, + "loss": 7.92384147644043, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.09320762753486633, + "learning_rate": 0.00026399999999999997, + "loss": 7.8478193283081055, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.09168332070112228, + "learning_rate": 0.00027, + "loss": 7.76185417175293, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.08832395076751709, + "learning_rate": 0.000276, + "loss": 7.679399490356445, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.08816400915384293, + "learning_rate": 0.00028199999999999997, + "loss": 7.5789642333984375, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0861717164516449, + "learning_rate": 0.00028799999999999995, + "loss": 7.519214630126953, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.08103746175765991, + "learning_rate": 0.000294, + "loss": 7.4475321769714355, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.07545104622840881, + "learning_rate": 0.0003, + "loss": 7.401182174682617, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.07042575627565384, + "learning_rate": 0.00030599999999999996, + "loss": 7.340173244476318, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.06737572699785233, + "learning_rate": 0.000312, + "loss": 7.265697002410889, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.06151146441698074, + "learning_rate": 0.000318, + "loss": 7.231061935424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.05638071522116661, + "learning_rate": 0.000324, + "loss": 7.195466995239258, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.05084463208913803, + "learning_rate": 0.00033, + "loss": 7.137603759765625, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.04293261095881462, + "learning_rate": 0.000336, + "loss": 7.115458965301514, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.036064863204956055, + "learning_rate": 0.00034199999999999996, + "loss": 7.089038848876953, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.029026754200458527, + "learning_rate": 0.00034799999999999995, + "loss": 7.0584306716918945, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.02286517433822155, + "learning_rate": 0.00035399999999999993, + "loss": 7.049450874328613, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.018878433853387833, + "learning_rate": 0.00035999999999999997, + "loss": 7.008824348449707, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.010773813351988792, + "learning_rate": 0.00036599999999999995, + "loss": 7.024829864501953, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.008794732391834259, + "learning_rate": 0.000372, + "loss": 7.018399238586426, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.011709796264767647, + "learning_rate": 0.00037799999999999997, + "loss": 7.0101752281188965, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.012123258784413338, + "learning_rate": 0.00038399999999999996, + "loss": 7.024317741394043, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01530320756137371, + "learning_rate": 0.00039, + "loss": 7.027738571166992, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.01890002004802227, + "learning_rate": 0.000396, + "loss": 7.024099349975586, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.020684150978922844, + "learning_rate": 0.000402, + "loss": 7.007943153381348, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.019826196134090424, + "learning_rate": 0.000408, + "loss": 7.006874084472656, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.019218673929572105, + "learning_rate": 0.0004139999999999999, + "loss": 7.002897262573242, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.013856411911547184, + "learning_rate": 0.00041999999999999996, + "loss": 6.993999481201172, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.010390003211796284, + "learning_rate": 0.00042599999999999995, + "loss": 7.000683784484863, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.91140079498291, + "eval_runtime": 57.7925, + "eval_samples_per_second": 42.255, + "eval_steps_per_second": 1.332, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.013802163302898407, + "learning_rate": 0.00043199999999999993, + "loss": 6.998257637023926, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.010109687224030495, + "learning_rate": 0.00043799999999999997, + "loss": 7.002542018890381, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.008019139990210533, + "learning_rate": 0.00044399999999999995, + "loss": 6.985941410064697, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.0069672465324401855, + "learning_rate": 0.00045, + "loss": 6.99245023727417, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.00714076729491353, + "learning_rate": 0.00045599999999999997, + "loss": 6.959296226501465, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.008408435620367527, + "learning_rate": 0.00046199999999999995, + "loss": 6.954504489898682, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.008778786286711693, + "learning_rate": 0.000468, + "loss": 7.0007004737854, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.007753407116979361, + "learning_rate": 0.000474, + "loss": 6.98183536529541, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.01087260153144598, + "learning_rate": 0.00047999999999999996, + "loss": 6.991066932678223, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.006122548598796129, + "learning_rate": 0.000486, + "loss": 6.978824138641357, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.0067383465357124805, + "learning_rate": 0.0004919999999999999, + "loss": 6.961703777313232, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.007922791875898838, + "learning_rate": 0.000498, + "loss": 6.954860687255859, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.00818922370672226, + "learning_rate": 0.0005039999999999999, + "loss": 6.945961952209473, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.006791461259126663, + "learning_rate": 0.0005099999999999999, + "loss": 6.968680381774902, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.005394163075834513, + "learning_rate": 0.000516, + "loss": 6.956404685974121, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.004500220064073801, + "learning_rate": 0.000522, + "loss": 6.96998405456543, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008699574507772923, + "learning_rate": 0.0005279999999999999, + "loss": 6.955585479736328, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.005889365915209055, + "learning_rate": 0.000534, + "loss": 6.9708051681518555, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009646810591220856, + "learning_rate": 0.00054, + "loss": 6.962608814239502, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.004517171997576952, + "learning_rate": 0.0005459999999999999, + "loss": 6.957627296447754, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.006339677143841982, + "learning_rate": 0.000552, + "loss": 6.963141441345215, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.005387383047491312, + "learning_rate": 0.000558, + "loss": 6.934465408325195, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.005252422299236059, + "learning_rate": 0.0005639999999999999, + "loss": 6.956108093261719, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.008699414320290089, + "learning_rate": 0.00057, + "loss": 6.973369598388672, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.003947409335523844, + "learning_rate": 0.0005759999999999999, + "loss": 6.950225830078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.00768681475892663, + "learning_rate": 0.0005819999999999999, + "loss": 6.964739799499512, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.006563829258084297, + "learning_rate": 0.000588, + "loss": 6.972421169281006, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007606555242091417, + "learning_rate": 0.0005939999999999999, + "loss": 6.958213806152344, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.006250880658626556, + "learning_rate": 0.0006, + "loss": 6.962296485900879, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.0073033408261835575, + "learning_rate": 0.0006, + "loss": 6.969705104827881, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.0054510910995304585, + "learning_rate": 0.0006, + "loss": 6.9652862548828125, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.004895673133432865, + "learning_rate": 0.0006, + "loss": 6.956023693084717, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.005306515377014875, + "learning_rate": 0.0006, + "loss": 6.963231086730957, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.004742146469652653, + "learning_rate": 0.0006, + "loss": 6.9582061767578125, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.004220000468194485, + "learning_rate": 0.0006, + "loss": 6.972961902618408, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0042153168469667435, + "learning_rate": 0.0006, + "loss": 6.961165904998779, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.005165040958672762, + "learning_rate": 0.0006, + "loss": 6.95018196105957, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.003930266480892897, + "learning_rate": 0.0006, + "loss": 6.9651689529418945, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.004465071018785238, + "learning_rate": 0.0006, + "loss": 6.960553169250488, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.004888612311333418, + "learning_rate": 0.0006, + "loss": 6.967258930206299, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.007497863378375769, + "learning_rate": 0.0006, + "loss": 6.954545021057129, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.005659648682922125, + "learning_rate": 0.0006, + "loss": 6.962299823760986, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.006641396787017584, + "learning_rate": 0.0006, + "loss": 6.952983856201172, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.009715928696095943, + "learning_rate": 0.0006, + "loss": 6.958622932434082, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.012789523229002953, + "learning_rate": 0.0006, + "loss": 6.932043075561523, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021673237904906273, + "learning_rate": 0.0006, + "loss": 6.931290149688721, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.16024991869926453, + "learning_rate": 0.0006, + "loss": 6.979497909545898, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.2434523105621338, + "learning_rate": 0.0006, + "loss": 7.185218811035156, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.39252421259880066, + "learning_rate": 0.0006, + "loss": 7.137226581573486, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.10565312206745148, + "learning_rate": 0.0006, + "loss": 6.984494209289551, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.1033431813120842, + "learning_rate": 0.0006, + "loss": 7.001287937164307, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.14875490963459015, + "learning_rate": 0.0006, + "loss": 7.051436901092529, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.1374562531709671, + "learning_rate": 0.0006, + "loss": 7.0293121337890625, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.0910211056470871, + "learning_rate": 0.0006, + "loss": 7.007458686828613, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.021633487194776535, + "learning_rate": 0.0006, + "loss": 6.996816635131836, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.08692754060029984, + "learning_rate": 0.0006, + "loss": 6.998541831970215, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.12613292038440704, + "learning_rate": 0.0006, + "loss": 6.9970703125, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.09462810307741165, + "learning_rate": 0.0006, + "loss": 6.997045040130615, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.03534962609410286, + "learning_rate": 0.0006, + "loss": 6.9738450050354, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.033839885145425797, + "learning_rate": 0.0006, + "loss": 6.946001052856445, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.06263269484043121, + "learning_rate": 0.0006, + "loss": 6.968981742858887, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07117997109889984, + "learning_rate": 0.0006, + "loss": 6.981535911560059, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.06336277723312378, + "learning_rate": 0.0006, + "loss": 6.976946830749512, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047467734664678574, + "learning_rate": 0.0006, + "loss": 6.961673736572266, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.018787868320941925, + "learning_rate": 0.0006, + "loss": 6.948919296264648, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.025909408926963806, + "learning_rate": 0.0006, + "loss": 6.976190567016602, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.052842333912849426, + "learning_rate": 0.0006, + "loss": 6.971207618713379, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.06421888619661331, + "learning_rate": 0.0006, + "loss": 6.979109764099121, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.055596910417079926, + "learning_rate": 0.0006, + "loss": 6.972118854522705, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.03267091140151024, + "learning_rate": 0.0006, + "loss": 6.947260856628418, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.008032613433897495, + "learning_rate": 0.0006, + "loss": 6.936816215515137, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.020184025168418884, + "learning_rate": 0.0006, + "loss": 6.929848670959473, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.836523056030273, + "eval_runtime": 58.5622, + "eval_samples_per_second": 41.699, + "eval_steps_per_second": 1.315, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.031980376690626144, + "learning_rate": 0.0006, + "loss": 6.968081474304199, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03802533075213432, + "learning_rate": 0.0006, + "loss": 6.95728874206543, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03796226158738136, + "learning_rate": 0.0006, + "loss": 6.96516752243042, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.02621951326727867, + "learning_rate": 0.0006, + "loss": 6.956058502197266, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.013292718678712845, + "learning_rate": 0.0006, + "loss": 6.9783935546875, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.009446968324482441, + "learning_rate": 0.0006, + "loss": 6.96873664855957, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.021703092381358147, + "learning_rate": 0.0006, + "loss": 6.959234237670898, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.027765290811657906, + "learning_rate": 0.0006, + "loss": 6.968524932861328, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.028761165216565132, + "learning_rate": 0.0006, + "loss": 6.954620838165283, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.019729383289813995, + "learning_rate": 0.0006, + "loss": 6.949009895324707, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.008990089409053326, + "learning_rate": 0.0006, + "loss": 6.956226348876953, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.00549644511193037, + "learning_rate": 0.0006, + "loss": 6.957143783569336, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.014436670579016209, + "learning_rate": 0.0006, + "loss": 6.954257011413574, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.01990494504570961, + "learning_rate": 0.0006, + "loss": 6.942131519317627, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.019155776128172874, + "learning_rate": 0.0006, + "loss": 6.946969985961914, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.014970783144235611, + "learning_rate": 0.0006, + "loss": 6.957938194274902, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.009875659830868244, + "learning_rate": 0.0006, + "loss": 6.953309059143066, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.004792087245732546, + "learning_rate": 0.0006, + "loss": 6.975146770477295, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.009936527349054813, + "learning_rate": 0.0006, + "loss": 6.961674690246582, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.015100127086043358, + "learning_rate": 0.0006, + "loss": 6.968759536743164, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01449542585760355, + "learning_rate": 0.0006, + "loss": 6.966510772705078, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.01073537115007639, + "learning_rate": 0.0006, + "loss": 6.936373710632324, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.006766272243112326, + "learning_rate": 0.0006, + "loss": 6.96397590637207, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.005055439658463001, + "learning_rate": 0.0006, + "loss": 6.961050987243652, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.006252622697502375, + "learning_rate": 0.0006, + "loss": 6.980506420135498, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.008084346540272236, + "learning_rate": 0.0006, + "loss": 6.9346604347229, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.010260380804538727, + "learning_rate": 0.0006, + "loss": 6.964118957519531, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.010697875171899796, + "learning_rate": 0.0006, + "loss": 6.941572189331055, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.007338830269873142, + "learning_rate": 0.0006, + "loss": 6.950471878051758, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.005931832827627659, + "learning_rate": 0.0006, + "loss": 6.962935924530029, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.0061243255622684956, + "learning_rate": 0.0006, + "loss": 6.946418285369873, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.007983328774571419, + "learning_rate": 0.0006, + "loss": 6.959758758544922, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.010470672510564327, + "learning_rate": 0.0006, + "loss": 6.93851375579834, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.011147388257086277, + "learning_rate": 0.0006, + "loss": 6.93258810043335, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01058544497936964, + "learning_rate": 0.0006, + "loss": 6.932259559631348, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.016850745305418968, + "learning_rate": 0.0006, + "loss": 6.924542427062988, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.023101547732949257, + "learning_rate": 0.0006, + "loss": 6.914878845214844, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.015795735642313957, + "learning_rate": 0.0006, + "loss": 6.922432899475098, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.008821426890790462, + "learning_rate": 0.0006, + "loss": 6.924562454223633, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.012502548284828663, + "learning_rate": 0.0006, + "loss": 6.914718151092529, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.011545066721737385, + "learning_rate": 0.0006, + "loss": 6.915705680847168, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.022190893068909645, + "learning_rate": 0.0006, + "loss": 6.88851261138916, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.023395851254463196, + "learning_rate": 0.0006, + "loss": 6.877655029296875, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015601156279444695, + "learning_rate": 0.0006, + "loss": 6.889644145965576, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01362934336066246, + "learning_rate": 0.0006, + "loss": 6.891547203063965, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.012305490672588348, + "learning_rate": 0.0006, + "loss": 6.890357971191406, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.01623358018696308, + "learning_rate": 0.0006, + "loss": 6.849443435668945, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.018063291907310486, + "learning_rate": 0.0006, + "loss": 6.842372894287109, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.010179434902966022, + "learning_rate": 0.0006, + "loss": 6.870055198669434, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.009986722841858864, + "learning_rate": 0.0006, + "loss": 6.859061241149902, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.01117654051631689, + "learning_rate": 0.0006, + "loss": 6.861199378967285, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.008551654405891895, + "learning_rate": 0.0006, + "loss": 6.850255012512207, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.018904326483607292, + "learning_rate": 0.0006, + "loss": 6.848639965057373, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.02080724574625492, + "learning_rate": 0.0006, + "loss": 6.80456018447876, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.012010873295366764, + "learning_rate": 0.0006, + "loss": 6.848775863647461, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.008113056421279907, + "learning_rate": 0.0006, + "loss": 6.849387168884277, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.012912592850625515, + "learning_rate": 0.0006, + "loss": 6.832240104675293, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.01146865077316761, + "learning_rate": 0.0006, + "loss": 6.850383281707764, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.0171892661601305, + "learning_rate": 0.0006, + "loss": 6.839577674865723, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.012083170004189014, + "learning_rate": 0.0006, + "loss": 6.818614482879639, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.0062555656768381596, + "learning_rate": 0.0006, + "loss": 6.831291198730469, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.010102466680109501, + "learning_rate": 0.0006, + "loss": 6.846522808074951, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.009042087942361832, + "learning_rate": 0.0006, + "loss": 6.822072982788086, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.011133556254208088, + "learning_rate": 0.0006, + "loss": 6.811252593994141, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.00990577507764101, + "learning_rate": 0.0006, + "loss": 6.814435958862305, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.006605073343962431, + "learning_rate": 0.0006, + "loss": 6.831300735473633, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.007831175811588764, + "learning_rate": 0.0006, + "loss": 6.841738700866699, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.005536255426704884, + "learning_rate": 0.0006, + "loss": 6.8043670654296875, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.006532667670398951, + "learning_rate": 0.0006, + "loss": 6.8091840744018555, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.006674565840512514, + "learning_rate": 0.0006, + "loss": 6.811327934265137, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.01150344219058752, + "learning_rate": 0.0006, + "loss": 6.810762405395508, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.009015528485178947, + "learning_rate": 0.0006, + "loss": 6.836431980133057, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.718832969665527, + "eval_runtime": 58.6615, + "eval_samples_per_second": 41.629, + "eval_steps_per_second": 1.313, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.009791829623281956, + "learning_rate": 0.0006, + "loss": 6.8102593421936035, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.008970947004854679, + "learning_rate": 0.0006, + "loss": 6.792663097381592, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.013689392246305943, + "learning_rate": 0.0006, + "loss": 6.815651893615723, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.031499531120061874, + "learning_rate": 0.0006, + "loss": 6.809841156005859, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03203094005584717, + "learning_rate": 0.0006, + "loss": 6.779417037963867, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.007947996258735657, + "learning_rate": 0.0006, + "loss": 6.773656845092773, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03152519091963768, + "learning_rate": 0.0006, + "loss": 6.804083824157715, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.014937258325517178, + "learning_rate": 0.0006, + "loss": 6.792213439941406, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.020961705595254898, + "learning_rate": 0.0006, + "loss": 6.8092803955078125, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.009152988903224468, + "learning_rate": 0.0006, + "loss": 6.7737603187561035, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.02770620957016945, + "learning_rate": 0.0006, + "loss": 6.767014503479004, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.01165871974080801, + "learning_rate": 0.0006, + "loss": 6.760884761810303, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.01910376362502575, + "learning_rate": 0.0006, + "loss": 6.775722980499268, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.010413747280836105, + "learning_rate": 0.0006, + "loss": 6.779816627502441, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.01918417401611805, + "learning_rate": 0.0006, + "loss": 6.760381698608398, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.017960334196686745, + "learning_rate": 0.0006, + "loss": 6.761198997497559, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.016546795144677162, + "learning_rate": 0.0006, + "loss": 6.742448806762695, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.01889597810804844, + "learning_rate": 0.0006, + "loss": 6.758237838745117, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.009611992165446281, + "learning_rate": 0.0006, + "loss": 6.75614070892334, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02034934051334858, + "learning_rate": 0.0006, + "loss": 6.736965179443359, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.010740065947175026, + "learning_rate": 0.0006, + "loss": 6.746947288513184, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.018383340910077095, + "learning_rate": 0.0006, + "loss": 6.747332572937012, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.012362911365926266, + "learning_rate": 0.0006, + "loss": 6.767669200897217, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.01127463299781084, + "learning_rate": 0.0006, + "loss": 6.769808292388916, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.017693769186735153, + "learning_rate": 0.0006, + "loss": 6.733658790588379, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.006420376244932413, + "learning_rate": 0.0006, + "loss": 6.74934720993042, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.014020364731550217, + "learning_rate": 0.0006, + "loss": 6.728157043457031, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.008018582127988338, + "learning_rate": 0.0006, + "loss": 6.730726718902588, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.00782222580164671, + "learning_rate": 0.0006, + "loss": 6.7499895095825195, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.008278409950435162, + "learning_rate": 0.0006, + "loss": 6.712803363800049, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.009826759807765484, + "learning_rate": 0.0006, + "loss": 6.742262840270996, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.009488224983215332, + "learning_rate": 0.0006, + "loss": 6.739130020141602, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.007667903788387775, + "learning_rate": 0.0006, + "loss": 6.747515678405762, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.0073090000078082085, + "learning_rate": 0.0006, + "loss": 6.740870475769043, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.010058708488941193, + "learning_rate": 0.0006, + "loss": 6.777563095092773, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.020676063373684883, + "learning_rate": 0.0006, + "loss": 6.708603858947754, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.027210544794797897, + "learning_rate": 0.0006, + "loss": 6.694097995758057, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.038053520023822784, + "learning_rate": 0.0006, + "loss": 6.697017669677734, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04684312269091606, + "learning_rate": 0.0006, + "loss": 6.729883193969727, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03673060983419418, + "learning_rate": 0.0006, + "loss": 6.743753433227539, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.006882632151246071, + "learning_rate": 0.0006, + "loss": 6.71405029296875, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.03588201105594635, + "learning_rate": 0.0006, + "loss": 6.712467193603516, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.029415199533104897, + "learning_rate": 0.0006, + "loss": 6.715634822845459, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.011425039730966091, + "learning_rate": 0.0006, + "loss": 6.697334289550781, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.0297263003885746, + "learning_rate": 0.0006, + "loss": 6.724848747253418, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.007282314356416464, + "learning_rate": 0.0006, + "loss": 6.702306270599365, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.02673051692545414, + "learning_rate": 0.0006, + "loss": 6.72357177734375, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.016174081712961197, + "learning_rate": 0.0006, + "loss": 6.730698585510254, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.0182416420429945, + "learning_rate": 0.0006, + "loss": 6.717386245727539, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.023250866681337357, + "learning_rate": 0.0006, + "loss": 6.697832107543945, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.007539753336459398, + "learning_rate": 0.0006, + "loss": 6.714348793029785, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.030086630955338478, + "learning_rate": 0.0006, + "loss": 6.724481582641602, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.012784913182258606, + "learning_rate": 0.0006, + "loss": 6.703491687774658, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.02042769268155098, + "learning_rate": 0.0006, + "loss": 6.721139907836914, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.023557329550385475, + "learning_rate": 0.0006, + "loss": 6.728122711181641, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.007716621737927198, + "learning_rate": 0.0006, + "loss": 6.70085334777832, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022772278636693954, + "learning_rate": 0.0006, + "loss": 6.712882995605469, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.007928933948278427, + "learning_rate": 0.0006, + "loss": 6.698320388793945, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.022355245426297188, + "learning_rate": 0.0006, + "loss": 6.687088489532471, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.02037850394845009, + "learning_rate": 0.0006, + "loss": 6.676856994628906, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.012290362268686295, + "learning_rate": 0.0006, + "loss": 6.682883262634277, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.022331275045871735, + "learning_rate": 0.0006, + "loss": 6.694752216339111, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.007527181413024664, + "learning_rate": 0.0006, + "loss": 6.686394691467285, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.02159409038722515, + "learning_rate": 0.0006, + "loss": 6.697943687438965, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.015054763294756413, + "learning_rate": 0.0006, + "loss": 6.677796363830566, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01466019731014967, + "learning_rate": 0.0006, + "loss": 6.697715759277344, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.01875465176999569, + "learning_rate": 0.0006, + "loss": 6.700228691101074, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.00778991961851716, + "learning_rate": 0.0006, + "loss": 6.69331693649292, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.020182441920042038, + "learning_rate": 0.0006, + "loss": 6.672651290893555, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.01240516733378172, + "learning_rate": 0.0006, + "loss": 6.688860893249512, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.011218412779271603, + "learning_rate": 0.0006, + "loss": 6.669534683227539, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.0141852255910635, + "learning_rate": 0.0006, + "loss": 6.690206527709961, + "step": 288 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.2278453445328896e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-288/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef +size 4792 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/config.json new file mode 100644 index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151674 +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0bf8f0e1914ebcf20f10ca6ef1d21e17f4e2fa9e --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea4335cd4b1048a8590c5723e07a74c2f88f7104372810a660a78b3e8a83153c +size 583362376 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..77daaeb55ca47e42103f5232718b3c16a7060a25 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:736618660f2f79bdb5b703a4de951220a9791c054c8458a3aad796bb4e4970f6 +size 1166837626 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..ac9172a69fff7bb676f49c9906b9314a59fdad2d --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c48c5ec5c9bb4c521c495daafd6b39a9c355c1b6424c2e78c7796e18f5fdd3fb +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..7f003aa3c9fce222cf1ebce0837bf5b92cde59e3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e8a93a2b7133c2fed28a7b289f0e7b9c9545ff71cee4aae92941f10b916a20f +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..86120748c38cd9ddec772def3d227c8f1512792b --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb930d5328a582540de3500ea4a3237c706f2698967029d50f2025f2c69fbf69 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..c35ff8c6ec57f5a32994c021011ed89dc2fb1637 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a71e181c472366aba18ce125f3793f840051870bcbc1acd0fa9ef87d62644fc2 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..dbc454a52c5d1edb241846f7cbf86268ced2c2ff --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f28b8022dab4a12fbbad2bccde532c0f7c60e469bde0bf66d40f7d043ac1727 +size 1064 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14 +size 11424648 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/tokenizer_config.json @@ -0,0 +1,21 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>", + "<|mask|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e15036434503c93bf7f8c05c5b04b17aa1ee6e5a --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/trainer_state.json @@ -0,0 +1,4122 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 8.0, + "eval_steps": 500, + "global_step": 576, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.5768774151802063, + "learning_rate": 0.0, + "loss": 12.050328254699707, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.5784570574760437, + "learning_rate": 5.999999999999999e-06, + "loss": 12.049251556396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.5607187151908875, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.921792030334473, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.46466919779777527, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.686628341674805, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.38012945652008057, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.432823181152344, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.33215776085853577, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.207542419433594, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.2928776741027832, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.000097274780273, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.26018962264060974, + "learning_rate": 4.2e-05, + "loss": 10.825748443603516, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.23717083036899567, + "learning_rate": 4.7999999999999994e-05, + "loss": 10.677122116088867, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.21641355752944946, + "learning_rate": 5.399999999999999e-05, + "loss": 10.569148063659668, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.1982956826686859, + "learning_rate": 5.9999999999999995e-05, + "loss": 10.49653434753418, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.18819282948970795, + "learning_rate": 6.599999999999999e-05, + "loss": 10.432951927185059, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.18322618305683136, + "learning_rate": 7.199999999999999e-05, + "loss": 10.387115478515625, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.18399226665496826, + "learning_rate": 7.8e-05, + "loss": 10.33381462097168, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.1833190619945526, + "learning_rate": 8.4e-05, + "loss": 10.301689147949219, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.18490834534168243, + "learning_rate": 8.999999999999999e-05, + "loss": 10.240968704223633, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.18358023464679718, + "learning_rate": 9.599999999999999e-05, + "loss": 10.194717407226562, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.1801682710647583, + "learning_rate": 0.000102, + "loss": 10.14897632598877, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.17899440228939056, + "learning_rate": 0.00010799999999999998, + "loss": 10.078763961791992, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.17666497826576233, + "learning_rate": 0.00011399999999999999, + "loss": 10.01095962524414, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.17196154594421387, + "learning_rate": 0.00011999999999999999, + "loss": 9.951203346252441, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.169203519821167, + "learning_rate": 0.00012599999999999997, + "loss": 9.871726036071777, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.16500306129455566, + "learning_rate": 0.00013199999999999998, + "loss": 9.803646087646484, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.16128124296665192, + "learning_rate": 0.000138, + "loss": 9.728487014770508, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.15904058516025543, + "learning_rate": 0.00014399999999999998, + "loss": 9.642473220825195, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.15560561418533325, + "learning_rate": 0.00015, + "loss": 9.56414794921875, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.15278612077236176, + "learning_rate": 0.000156, + "loss": 9.484230041503906, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.15263237059116364, + "learning_rate": 0.000162, + "loss": 9.383644104003906, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.14945024251937866, + "learning_rate": 0.000168, + "loss": 9.29840087890625, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.14614646136760712, + "learning_rate": 0.00017399999999999997, + "loss": 9.208931922912598, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.14295098185539246, + "learning_rate": 0.00017999999999999998, + "loss": 9.122323989868164, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.13868345320224762, + "learning_rate": 0.000186, + "loss": 9.03034496307373, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.135064035654068, + "learning_rate": 0.00019199999999999998, + "loss": 8.92818832397461, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.13097241520881653, + "learning_rate": 0.000198, + "loss": 8.834627151489258, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.12478776276111603, + "learning_rate": 0.000204, + "loss": 8.755172729492188, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.1214982345700264, + "learning_rate": 0.00020999999999999998, + "loss": 8.647891998291016, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.11958328634500504, + "learning_rate": 0.00021599999999999996, + "loss": 8.543274879455566, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.11680915951728821, + "learning_rate": 0.00022199999999999998, + "loss": 8.44540023803711, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.11491474509239197, + "learning_rate": 0.00022799999999999999, + "loss": 8.349921226501465, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.11160755902528763, + "learning_rate": 0.000234, + "loss": 8.26960563659668, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.10840637236833572, + "learning_rate": 0.00023999999999999998, + "loss": 8.192136764526367, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.10585298389196396, + "learning_rate": 0.00024599999999999996, + "loss": 8.09546947479248, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.1032542735338211, + "learning_rate": 0.00025199999999999995, + "loss": 7.991090774536133, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.0986943170428276, + "learning_rate": 0.000258, + "loss": 7.92384147644043, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.09320762753486633, + "learning_rate": 0.00026399999999999997, + "loss": 7.8478193283081055, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.09168332070112228, + "learning_rate": 0.00027, + "loss": 7.76185417175293, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.08832395076751709, + "learning_rate": 0.000276, + "loss": 7.679399490356445, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.08816400915384293, + "learning_rate": 0.00028199999999999997, + "loss": 7.5789642333984375, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0861717164516449, + "learning_rate": 0.00028799999999999995, + "loss": 7.519214630126953, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.08103746175765991, + "learning_rate": 0.000294, + "loss": 7.4475321769714355, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.07545104622840881, + "learning_rate": 0.0003, + "loss": 7.401182174682617, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.07042575627565384, + "learning_rate": 0.00030599999999999996, + "loss": 7.340173244476318, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.06737572699785233, + "learning_rate": 0.000312, + "loss": 7.265697002410889, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.06151146441698074, + "learning_rate": 0.000318, + "loss": 7.231061935424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.05638071522116661, + "learning_rate": 0.000324, + "loss": 7.195466995239258, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.05084463208913803, + "learning_rate": 0.00033, + "loss": 7.137603759765625, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.04293261095881462, + "learning_rate": 0.000336, + "loss": 7.115458965301514, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.036064863204956055, + "learning_rate": 0.00034199999999999996, + "loss": 7.089038848876953, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.029026754200458527, + "learning_rate": 0.00034799999999999995, + "loss": 7.0584306716918945, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.02286517433822155, + "learning_rate": 0.00035399999999999993, + "loss": 7.049450874328613, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.018878433853387833, + "learning_rate": 0.00035999999999999997, + "loss": 7.008824348449707, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.010773813351988792, + "learning_rate": 0.00036599999999999995, + "loss": 7.024829864501953, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.008794732391834259, + "learning_rate": 0.000372, + "loss": 7.018399238586426, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.011709796264767647, + "learning_rate": 0.00037799999999999997, + "loss": 7.0101752281188965, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.012123258784413338, + "learning_rate": 0.00038399999999999996, + "loss": 7.024317741394043, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01530320756137371, + "learning_rate": 0.00039, + "loss": 7.027738571166992, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.01890002004802227, + "learning_rate": 0.000396, + "loss": 7.024099349975586, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.020684150978922844, + "learning_rate": 0.000402, + "loss": 7.007943153381348, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.019826196134090424, + "learning_rate": 0.000408, + "loss": 7.006874084472656, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.019218673929572105, + "learning_rate": 0.0004139999999999999, + "loss": 7.002897262573242, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.013856411911547184, + "learning_rate": 0.00041999999999999996, + "loss": 6.993999481201172, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.010390003211796284, + "learning_rate": 0.00042599999999999995, + "loss": 7.000683784484863, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.91140079498291, + "eval_runtime": 57.7925, + "eval_samples_per_second": 42.255, + "eval_steps_per_second": 1.332, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.013802163302898407, + "learning_rate": 0.00043199999999999993, + "loss": 6.998257637023926, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.010109687224030495, + "learning_rate": 0.00043799999999999997, + "loss": 7.002542018890381, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.008019139990210533, + "learning_rate": 0.00044399999999999995, + "loss": 6.985941410064697, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.0069672465324401855, + "learning_rate": 0.00045, + "loss": 6.99245023727417, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.00714076729491353, + "learning_rate": 0.00045599999999999997, + "loss": 6.959296226501465, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.008408435620367527, + "learning_rate": 0.00046199999999999995, + "loss": 6.954504489898682, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.008778786286711693, + "learning_rate": 0.000468, + "loss": 7.0007004737854, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.007753407116979361, + "learning_rate": 0.000474, + "loss": 6.98183536529541, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.01087260153144598, + "learning_rate": 0.00047999999999999996, + "loss": 6.991066932678223, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.006122548598796129, + "learning_rate": 0.000486, + "loss": 6.978824138641357, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.0067383465357124805, + "learning_rate": 0.0004919999999999999, + "loss": 6.961703777313232, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.007922791875898838, + "learning_rate": 0.000498, + "loss": 6.954860687255859, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.00818922370672226, + "learning_rate": 0.0005039999999999999, + "loss": 6.945961952209473, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.006791461259126663, + "learning_rate": 0.0005099999999999999, + "loss": 6.968680381774902, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.005394163075834513, + "learning_rate": 0.000516, + "loss": 6.956404685974121, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.004500220064073801, + "learning_rate": 0.000522, + "loss": 6.96998405456543, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008699574507772923, + "learning_rate": 0.0005279999999999999, + "loss": 6.955585479736328, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.005889365915209055, + "learning_rate": 0.000534, + "loss": 6.9708051681518555, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009646810591220856, + "learning_rate": 0.00054, + "loss": 6.962608814239502, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.004517171997576952, + "learning_rate": 0.0005459999999999999, + "loss": 6.957627296447754, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.006339677143841982, + "learning_rate": 0.000552, + "loss": 6.963141441345215, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.005387383047491312, + "learning_rate": 0.000558, + "loss": 6.934465408325195, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.005252422299236059, + "learning_rate": 0.0005639999999999999, + "loss": 6.956108093261719, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.008699414320290089, + "learning_rate": 0.00057, + "loss": 6.973369598388672, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.003947409335523844, + "learning_rate": 0.0005759999999999999, + "loss": 6.950225830078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.00768681475892663, + "learning_rate": 0.0005819999999999999, + "loss": 6.964739799499512, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.006563829258084297, + "learning_rate": 0.000588, + "loss": 6.972421169281006, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007606555242091417, + "learning_rate": 0.0005939999999999999, + "loss": 6.958213806152344, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.006250880658626556, + "learning_rate": 0.0006, + "loss": 6.962296485900879, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.0073033408261835575, + "learning_rate": 0.0006, + "loss": 6.969705104827881, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.0054510910995304585, + "learning_rate": 0.0006, + "loss": 6.9652862548828125, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.004895673133432865, + "learning_rate": 0.0006, + "loss": 6.956023693084717, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.005306515377014875, + "learning_rate": 0.0006, + "loss": 6.963231086730957, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.004742146469652653, + "learning_rate": 0.0006, + "loss": 6.9582061767578125, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.004220000468194485, + "learning_rate": 0.0006, + "loss": 6.972961902618408, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0042153168469667435, + "learning_rate": 0.0006, + "loss": 6.961165904998779, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.005165040958672762, + "learning_rate": 0.0006, + "loss": 6.95018196105957, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.003930266480892897, + "learning_rate": 0.0006, + "loss": 6.9651689529418945, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.004465071018785238, + "learning_rate": 0.0006, + "loss": 6.960553169250488, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.004888612311333418, + "learning_rate": 0.0006, + "loss": 6.967258930206299, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.007497863378375769, + "learning_rate": 0.0006, + "loss": 6.954545021057129, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.005659648682922125, + "learning_rate": 0.0006, + "loss": 6.962299823760986, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.006641396787017584, + "learning_rate": 0.0006, + "loss": 6.952983856201172, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.009715928696095943, + "learning_rate": 0.0006, + "loss": 6.958622932434082, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.012789523229002953, + "learning_rate": 0.0006, + "loss": 6.932043075561523, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021673237904906273, + "learning_rate": 0.0006, + "loss": 6.931290149688721, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.16024991869926453, + "learning_rate": 0.0006, + "loss": 6.979497909545898, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.2434523105621338, + "learning_rate": 0.0006, + "loss": 7.185218811035156, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.39252421259880066, + "learning_rate": 0.0006, + "loss": 7.137226581573486, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.10565312206745148, + "learning_rate": 0.0006, + "loss": 6.984494209289551, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.1033431813120842, + "learning_rate": 0.0006, + "loss": 7.001287937164307, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.14875490963459015, + "learning_rate": 0.0006, + "loss": 7.051436901092529, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.1374562531709671, + "learning_rate": 0.0006, + "loss": 7.0293121337890625, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.0910211056470871, + "learning_rate": 0.0006, + "loss": 7.007458686828613, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.021633487194776535, + "learning_rate": 0.0006, + "loss": 6.996816635131836, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.08692754060029984, + "learning_rate": 0.0006, + "loss": 6.998541831970215, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.12613292038440704, + "learning_rate": 0.0006, + "loss": 6.9970703125, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.09462810307741165, + "learning_rate": 0.0006, + "loss": 6.997045040130615, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.03534962609410286, + "learning_rate": 0.0006, + "loss": 6.9738450050354, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.033839885145425797, + "learning_rate": 0.0006, + "loss": 6.946001052856445, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.06263269484043121, + "learning_rate": 0.0006, + "loss": 6.968981742858887, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07117997109889984, + "learning_rate": 0.0006, + "loss": 6.981535911560059, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.06336277723312378, + "learning_rate": 0.0006, + "loss": 6.976946830749512, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047467734664678574, + "learning_rate": 0.0006, + "loss": 6.961673736572266, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.018787868320941925, + "learning_rate": 0.0006, + "loss": 6.948919296264648, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.025909408926963806, + "learning_rate": 0.0006, + "loss": 6.976190567016602, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.052842333912849426, + "learning_rate": 0.0006, + "loss": 6.971207618713379, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.06421888619661331, + "learning_rate": 0.0006, + "loss": 6.979109764099121, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.055596910417079926, + "learning_rate": 0.0006, + "loss": 6.972118854522705, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.03267091140151024, + "learning_rate": 0.0006, + "loss": 6.947260856628418, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.008032613433897495, + "learning_rate": 0.0006, + "loss": 6.936816215515137, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.020184025168418884, + "learning_rate": 0.0006, + "loss": 6.929848670959473, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.836523056030273, + "eval_runtime": 58.5622, + "eval_samples_per_second": 41.699, + "eval_steps_per_second": 1.315, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.031980376690626144, + "learning_rate": 0.0006, + "loss": 6.968081474304199, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03802533075213432, + "learning_rate": 0.0006, + "loss": 6.95728874206543, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03796226158738136, + "learning_rate": 0.0006, + "loss": 6.96516752243042, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.02621951326727867, + "learning_rate": 0.0006, + "loss": 6.956058502197266, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.013292718678712845, + "learning_rate": 0.0006, + "loss": 6.9783935546875, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.009446968324482441, + "learning_rate": 0.0006, + "loss": 6.96873664855957, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.021703092381358147, + "learning_rate": 0.0006, + "loss": 6.959234237670898, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.027765290811657906, + "learning_rate": 0.0006, + "loss": 6.968524932861328, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.028761165216565132, + "learning_rate": 0.0006, + "loss": 6.954620838165283, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.019729383289813995, + "learning_rate": 0.0006, + "loss": 6.949009895324707, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.008990089409053326, + "learning_rate": 0.0006, + "loss": 6.956226348876953, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.00549644511193037, + "learning_rate": 0.0006, + "loss": 6.957143783569336, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.014436670579016209, + "learning_rate": 0.0006, + "loss": 6.954257011413574, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.01990494504570961, + "learning_rate": 0.0006, + "loss": 6.942131519317627, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.019155776128172874, + "learning_rate": 0.0006, + "loss": 6.946969985961914, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.014970783144235611, + "learning_rate": 0.0006, + "loss": 6.957938194274902, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.009875659830868244, + "learning_rate": 0.0006, + "loss": 6.953309059143066, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.004792087245732546, + "learning_rate": 0.0006, + "loss": 6.975146770477295, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.009936527349054813, + "learning_rate": 0.0006, + "loss": 6.961674690246582, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.015100127086043358, + "learning_rate": 0.0006, + "loss": 6.968759536743164, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01449542585760355, + "learning_rate": 0.0006, + "loss": 6.966510772705078, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.01073537115007639, + "learning_rate": 0.0006, + "loss": 6.936373710632324, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.006766272243112326, + "learning_rate": 0.0006, + "loss": 6.96397590637207, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.005055439658463001, + "learning_rate": 0.0006, + "loss": 6.961050987243652, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.006252622697502375, + "learning_rate": 0.0006, + "loss": 6.980506420135498, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.008084346540272236, + "learning_rate": 0.0006, + "loss": 6.9346604347229, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.010260380804538727, + "learning_rate": 0.0006, + "loss": 6.964118957519531, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.010697875171899796, + "learning_rate": 0.0006, + "loss": 6.941572189331055, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.007338830269873142, + "learning_rate": 0.0006, + "loss": 6.950471878051758, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.005931832827627659, + "learning_rate": 0.0006, + "loss": 6.962935924530029, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.0061243255622684956, + "learning_rate": 0.0006, + "loss": 6.946418285369873, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.007983328774571419, + "learning_rate": 0.0006, + "loss": 6.959758758544922, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.010470672510564327, + "learning_rate": 0.0006, + "loss": 6.93851375579834, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.011147388257086277, + "learning_rate": 0.0006, + "loss": 6.93258810043335, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01058544497936964, + "learning_rate": 0.0006, + "loss": 6.932259559631348, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.016850745305418968, + "learning_rate": 0.0006, + "loss": 6.924542427062988, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.023101547732949257, + "learning_rate": 0.0006, + "loss": 6.914878845214844, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.015795735642313957, + "learning_rate": 0.0006, + "loss": 6.922432899475098, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.008821426890790462, + "learning_rate": 0.0006, + "loss": 6.924562454223633, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.012502548284828663, + "learning_rate": 0.0006, + "loss": 6.914718151092529, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.011545066721737385, + "learning_rate": 0.0006, + "loss": 6.915705680847168, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.022190893068909645, + "learning_rate": 0.0006, + "loss": 6.88851261138916, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.023395851254463196, + "learning_rate": 0.0006, + "loss": 6.877655029296875, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015601156279444695, + "learning_rate": 0.0006, + "loss": 6.889644145965576, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01362934336066246, + "learning_rate": 0.0006, + "loss": 6.891547203063965, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.012305490672588348, + "learning_rate": 0.0006, + "loss": 6.890357971191406, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.01623358018696308, + "learning_rate": 0.0006, + "loss": 6.849443435668945, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.018063291907310486, + "learning_rate": 0.0006, + "loss": 6.842372894287109, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.010179434902966022, + "learning_rate": 0.0006, + "loss": 6.870055198669434, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.009986722841858864, + "learning_rate": 0.0006, + "loss": 6.859061241149902, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.01117654051631689, + "learning_rate": 0.0006, + "loss": 6.861199378967285, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.008551654405891895, + "learning_rate": 0.0006, + "loss": 6.850255012512207, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.018904326483607292, + "learning_rate": 0.0006, + "loss": 6.848639965057373, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.02080724574625492, + "learning_rate": 0.0006, + "loss": 6.80456018447876, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.012010873295366764, + "learning_rate": 0.0006, + "loss": 6.848775863647461, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.008113056421279907, + "learning_rate": 0.0006, + "loss": 6.849387168884277, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.012912592850625515, + "learning_rate": 0.0006, + "loss": 6.832240104675293, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.01146865077316761, + "learning_rate": 0.0006, + "loss": 6.850383281707764, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.0171892661601305, + "learning_rate": 0.0006, + "loss": 6.839577674865723, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.012083170004189014, + "learning_rate": 0.0006, + "loss": 6.818614482879639, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.0062555656768381596, + "learning_rate": 0.0006, + "loss": 6.831291198730469, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.010102466680109501, + "learning_rate": 0.0006, + "loss": 6.846522808074951, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.009042087942361832, + "learning_rate": 0.0006, + "loss": 6.822072982788086, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.011133556254208088, + "learning_rate": 0.0006, + "loss": 6.811252593994141, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.00990577507764101, + "learning_rate": 0.0006, + "loss": 6.814435958862305, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.006605073343962431, + "learning_rate": 0.0006, + "loss": 6.831300735473633, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.007831175811588764, + "learning_rate": 0.0006, + "loss": 6.841738700866699, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.005536255426704884, + "learning_rate": 0.0006, + "loss": 6.8043670654296875, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.006532667670398951, + "learning_rate": 0.0006, + "loss": 6.8091840744018555, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.006674565840512514, + "learning_rate": 0.0006, + "loss": 6.811327934265137, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.01150344219058752, + "learning_rate": 0.0006, + "loss": 6.810762405395508, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.009015528485178947, + "learning_rate": 0.0006, + "loss": 6.836431980133057, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.718832969665527, + "eval_runtime": 58.6615, + "eval_samples_per_second": 41.629, + "eval_steps_per_second": 1.313, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.009791829623281956, + "learning_rate": 0.0006, + "loss": 6.8102593421936035, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.008970947004854679, + "learning_rate": 0.0006, + "loss": 6.792663097381592, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.013689392246305943, + "learning_rate": 0.0006, + "loss": 6.815651893615723, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.031499531120061874, + "learning_rate": 0.0006, + "loss": 6.809841156005859, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03203094005584717, + "learning_rate": 0.0006, + "loss": 6.779417037963867, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.007947996258735657, + "learning_rate": 0.0006, + "loss": 6.773656845092773, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03152519091963768, + "learning_rate": 0.0006, + "loss": 6.804083824157715, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.014937258325517178, + "learning_rate": 0.0006, + "loss": 6.792213439941406, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.020961705595254898, + "learning_rate": 0.0006, + "loss": 6.8092803955078125, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.009152988903224468, + "learning_rate": 0.0006, + "loss": 6.7737603187561035, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.02770620957016945, + "learning_rate": 0.0006, + "loss": 6.767014503479004, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.01165871974080801, + "learning_rate": 0.0006, + "loss": 6.760884761810303, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.01910376362502575, + "learning_rate": 0.0006, + "loss": 6.775722980499268, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.010413747280836105, + "learning_rate": 0.0006, + "loss": 6.779816627502441, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.01918417401611805, + "learning_rate": 0.0006, + "loss": 6.760381698608398, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.017960334196686745, + "learning_rate": 0.0006, + "loss": 6.761198997497559, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.016546795144677162, + "learning_rate": 0.0006, + "loss": 6.742448806762695, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.01889597810804844, + "learning_rate": 0.0006, + "loss": 6.758237838745117, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.009611992165446281, + "learning_rate": 0.0006, + "loss": 6.75614070892334, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02034934051334858, + "learning_rate": 0.0006, + "loss": 6.736965179443359, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.010740065947175026, + "learning_rate": 0.0006, + "loss": 6.746947288513184, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.018383340910077095, + "learning_rate": 0.0006, + "loss": 6.747332572937012, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.012362911365926266, + "learning_rate": 0.0006, + "loss": 6.767669200897217, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.01127463299781084, + "learning_rate": 0.0006, + "loss": 6.769808292388916, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.017693769186735153, + "learning_rate": 0.0006, + "loss": 6.733658790588379, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.006420376244932413, + "learning_rate": 0.0006, + "loss": 6.74934720993042, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.014020364731550217, + "learning_rate": 0.0006, + "loss": 6.728157043457031, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.008018582127988338, + "learning_rate": 0.0006, + "loss": 6.730726718902588, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.00782222580164671, + "learning_rate": 0.0006, + "loss": 6.7499895095825195, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.008278409950435162, + "learning_rate": 0.0006, + "loss": 6.712803363800049, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.009826759807765484, + "learning_rate": 0.0006, + "loss": 6.742262840270996, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.009488224983215332, + "learning_rate": 0.0006, + "loss": 6.739130020141602, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.007667903788387775, + "learning_rate": 0.0006, + "loss": 6.747515678405762, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.0073090000078082085, + "learning_rate": 0.0006, + "loss": 6.740870475769043, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.010058708488941193, + "learning_rate": 0.0006, + "loss": 6.777563095092773, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.020676063373684883, + "learning_rate": 0.0006, + "loss": 6.708603858947754, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.027210544794797897, + "learning_rate": 0.0006, + "loss": 6.694097995758057, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.038053520023822784, + "learning_rate": 0.0006, + "loss": 6.697017669677734, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04684312269091606, + "learning_rate": 0.0006, + "loss": 6.729883193969727, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03673060983419418, + "learning_rate": 0.0006, + "loss": 6.743753433227539, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.006882632151246071, + "learning_rate": 0.0006, + "loss": 6.71405029296875, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.03588201105594635, + "learning_rate": 0.0006, + "loss": 6.712467193603516, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.029415199533104897, + "learning_rate": 0.0006, + "loss": 6.715634822845459, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.011425039730966091, + "learning_rate": 0.0006, + "loss": 6.697334289550781, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.0297263003885746, + "learning_rate": 0.0006, + "loss": 6.724848747253418, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.007282314356416464, + "learning_rate": 0.0006, + "loss": 6.702306270599365, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.02673051692545414, + "learning_rate": 0.0006, + "loss": 6.72357177734375, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.016174081712961197, + "learning_rate": 0.0006, + "loss": 6.730698585510254, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.0182416420429945, + "learning_rate": 0.0006, + "loss": 6.717386245727539, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.023250866681337357, + "learning_rate": 0.0006, + "loss": 6.697832107543945, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.007539753336459398, + "learning_rate": 0.0006, + "loss": 6.714348793029785, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.030086630955338478, + "learning_rate": 0.0006, + "loss": 6.724481582641602, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.012784913182258606, + "learning_rate": 0.0006, + "loss": 6.703491687774658, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.02042769268155098, + "learning_rate": 0.0006, + "loss": 6.721139907836914, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.023557329550385475, + "learning_rate": 0.0006, + "loss": 6.728122711181641, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.007716621737927198, + "learning_rate": 0.0006, + "loss": 6.70085334777832, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022772278636693954, + "learning_rate": 0.0006, + "loss": 6.712882995605469, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.007928933948278427, + "learning_rate": 0.0006, + "loss": 6.698320388793945, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.022355245426297188, + "learning_rate": 0.0006, + "loss": 6.687088489532471, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.02037850394845009, + "learning_rate": 0.0006, + "loss": 6.676856994628906, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.012290362268686295, + "learning_rate": 0.0006, + "loss": 6.682883262634277, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.022331275045871735, + "learning_rate": 0.0006, + "loss": 6.694752216339111, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.007527181413024664, + "learning_rate": 0.0006, + "loss": 6.686394691467285, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.02159409038722515, + "learning_rate": 0.0006, + "loss": 6.697943687438965, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.015054763294756413, + "learning_rate": 0.0006, + "loss": 6.677796363830566, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01466019731014967, + "learning_rate": 0.0006, + "loss": 6.697715759277344, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.01875465176999569, + "learning_rate": 0.0006, + "loss": 6.700228691101074, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.00778991961851716, + "learning_rate": 0.0006, + "loss": 6.69331693649292, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.020182441920042038, + "learning_rate": 0.0006, + "loss": 6.672651290893555, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.01240516733378172, + "learning_rate": 0.0006, + "loss": 6.688860893249512, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.011218412779271603, + "learning_rate": 0.0006, + "loss": 6.669534683227539, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.0141852255910635, + "learning_rate": 0.0006, + "loss": 6.690206527709961, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 7.537620544433594, + "eval_runtime": 58.4398, + "eval_samples_per_second": 41.787, + "eval_steps_per_second": 1.318, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.008106146939098835, + "learning_rate": 0.0006, + "loss": 6.688591003417969, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.014588244259357452, + "learning_rate": 0.0006, + "loss": 6.684961318969727, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.011985368095338345, + "learning_rate": 0.0006, + "loss": 6.672794342041016, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.014632102102041245, + "learning_rate": 0.0006, + "loss": 6.702736854553223, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.011243713088333607, + "learning_rate": 0.0006, + "loss": 6.658792495727539, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.010519202798604965, + "learning_rate": 0.0006, + "loss": 6.662495136260986, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.007318088319152594, + "learning_rate": 0.0006, + "loss": 6.675954818725586, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.011712398380041122, + "learning_rate": 0.0006, + "loss": 6.673973083496094, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.006726635154336691, + "learning_rate": 0.0006, + "loss": 6.656833171844482, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.005466983187943697, + "learning_rate": 0.0006, + "loss": 6.675468921661377, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.007163621485233307, + "learning_rate": 0.0006, + "loss": 6.6862874031066895, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.007626629434525967, + "learning_rate": 0.0006, + "loss": 6.654741287231445, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.0066330647096037865, + "learning_rate": 0.0006, + "loss": 6.68114709854126, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.007363482844084501, + "learning_rate": 0.0006, + "loss": 6.683980941772461, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.00971233006566763, + "learning_rate": 0.0006, + "loss": 6.669700622558594, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.013185256160795689, + "learning_rate": 0.0006, + "loss": 6.686490535736084, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.019116027280688286, + "learning_rate": 0.0006, + "loss": 6.645479202270508, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0328814722597599, + "learning_rate": 0.0006, + "loss": 6.68080997467041, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04482287913560867, + "learning_rate": 0.0006, + "loss": 6.698016166687012, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.03053313121199608, + "learning_rate": 0.0006, + "loss": 6.642240524291992, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.043351784348487854, + "learning_rate": 0.0006, + "loss": 6.662343978881836, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.03165886923670769, + "learning_rate": 0.0006, + "loss": 6.6291961669921875, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.028601787984371185, + "learning_rate": 0.0006, + "loss": 6.667544364929199, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.019558480009436607, + "learning_rate": 0.0006, + "loss": 6.641025543212891, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.039851363748311996, + "learning_rate": 0.0006, + "loss": 6.665952205657959, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.018417146056890488, + "learning_rate": 0.0006, + "loss": 6.657257556915283, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.02894527092576027, + "learning_rate": 0.0006, + "loss": 6.656288146972656, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.01910966821014881, + "learning_rate": 0.0006, + "loss": 6.688958168029785, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.01668412610888481, + "learning_rate": 0.0006, + "loss": 6.64876127243042, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02935275249183178, + "learning_rate": 0.0006, + "loss": 6.626096725463867, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.017997972667217255, + "learning_rate": 0.0006, + "loss": 6.640386581420898, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.01779189519584179, + "learning_rate": 0.0006, + "loss": 6.641165256500244, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.013901712372899055, + "learning_rate": 0.0006, + "loss": 6.651627063751221, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013211199082434177, + "learning_rate": 0.0006, + "loss": 6.630064964294434, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.02324107475578785, + "learning_rate": 0.0006, + "loss": 6.670443058013916, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.019452504813671112, + "learning_rate": 0.0006, + "loss": 6.651266098022461, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.017943572252988815, + "learning_rate": 0.0006, + "loss": 6.657943248748779, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.012136237695813179, + "learning_rate": 0.0006, + "loss": 6.640082359313965, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.011544011533260345, + "learning_rate": 0.0006, + "loss": 6.6186933517456055, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01569426618516445, + "learning_rate": 0.0006, + "loss": 6.640767574310303, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.027668394148349762, + "learning_rate": 0.0006, + "loss": 6.651412487030029, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.04073230177164078, + "learning_rate": 0.0006, + "loss": 6.636565208435059, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.04942672327160835, + "learning_rate": 0.0006, + "loss": 6.667994499206543, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.037568598985672, + "learning_rate": 0.0006, + "loss": 6.632008075714111, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.009880400262773037, + "learning_rate": 0.0006, + "loss": 6.6504058837890625, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.033384259790182114, + "learning_rate": 0.0006, + "loss": 6.63852596282959, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.035640884190797806, + "learning_rate": 0.0006, + "loss": 6.627986431121826, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.013637864030897617, + "learning_rate": 0.0006, + "loss": 6.624503135681152, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.025304675102233887, + "learning_rate": 0.0006, + "loss": 6.612077236175537, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.03377160802483559, + "learning_rate": 0.0006, + "loss": 6.608938217163086, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.019803153350949287, + "learning_rate": 0.0006, + "loss": 6.636343955993652, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.018928129225969315, + "learning_rate": 0.0006, + "loss": 6.636613845825195, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.023068996146321297, + "learning_rate": 0.0006, + "loss": 6.6258440017700195, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.016849957406520844, + "learning_rate": 0.0006, + "loss": 6.638483047485352, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.021430544555187225, + "learning_rate": 0.0006, + "loss": 6.619649410247803, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.024053368717432022, + "learning_rate": 0.0006, + "loss": 6.639214515686035, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.011926693841814995, + "learning_rate": 0.0006, + "loss": 6.604836463928223, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01961611397564411, + "learning_rate": 0.0006, + "loss": 6.624864101409912, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.027946067973971367, + "learning_rate": 0.0006, + "loss": 6.621284484863281, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.02108592912554741, + "learning_rate": 0.0006, + "loss": 6.5982465744018555, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.017272260040044785, + "learning_rate": 0.0006, + "loss": 6.626142501831055, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.015654582530260086, + "learning_rate": 0.0006, + "loss": 6.621735572814941, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.015797540545463562, + "learning_rate": 0.0006, + "loss": 6.6016411781311035, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.028488904237747192, + "learning_rate": 0.0006, + "loss": 6.598413467407227, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.03411734476685524, + "learning_rate": 0.0006, + "loss": 6.597321033477783, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.027188263833522797, + "learning_rate": 0.0006, + "loss": 6.630616188049316, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.03411437198519707, + "learning_rate": 0.0006, + "loss": 6.57647705078125, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.019907327368855476, + "learning_rate": 0.0006, + "loss": 6.579545974731445, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.02485930174589157, + "learning_rate": 0.0006, + "loss": 6.591073036193848, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.026106052100658417, + "learning_rate": 0.0006, + "loss": 6.580729961395264, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.029401391744613647, + "learning_rate": 0.0006, + "loss": 6.5460076332092285, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.032869379967451096, + "learning_rate": 0.0006, + "loss": 6.583278179168701, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 7.416288375854492, + "eval_runtime": 60.4177, + "eval_samples_per_second": 40.419, + "eval_steps_per_second": 1.274, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.04862433671951294, + "learning_rate": 0.0006, + "loss": 6.58473014831543, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.06636223942041397, + "learning_rate": 0.0006, + "loss": 6.621153354644775, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.05888264626264572, + "learning_rate": 0.0006, + "loss": 6.621196269989014, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.05331697314977646, + "learning_rate": 0.0006, + "loss": 6.600763320922852, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03947671875357628, + "learning_rate": 0.0006, + "loss": 6.600671768188477, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.04045343026518822, + "learning_rate": 0.0006, + "loss": 6.63529109954834, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.029766036197543144, + "learning_rate": 0.0006, + "loss": 6.612868309020996, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.03416327014565468, + "learning_rate": 0.0006, + "loss": 6.582240581512451, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03458515554666519, + "learning_rate": 0.0006, + "loss": 6.604381561279297, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.024099089205265045, + "learning_rate": 0.0006, + "loss": 6.597355365753174, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.02694438211619854, + "learning_rate": 0.0006, + "loss": 6.597512245178223, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024201856926083565, + "learning_rate": 0.0006, + "loss": 6.564685821533203, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.015959380194544792, + "learning_rate": 0.0006, + "loss": 6.582263946533203, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.025690171867609024, + "learning_rate": 0.0006, + "loss": 6.601057052612305, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.014497052878141403, + "learning_rate": 0.0006, + "loss": 6.584796905517578, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.019870297983288765, + "learning_rate": 0.0006, + "loss": 6.574010372161865, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.01807512901723385, + "learning_rate": 0.0006, + "loss": 6.558596134185791, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.014525193721055984, + "learning_rate": 0.0006, + "loss": 6.571803092956543, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.016113435849547386, + "learning_rate": 0.0006, + "loss": 6.568970680236816, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020724017173051834, + "learning_rate": 0.0006, + "loss": 6.5669403076171875, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.013963675126433372, + "learning_rate": 0.0006, + "loss": 6.586382865905762, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.01621115580201149, + "learning_rate": 0.0006, + "loss": 6.539546966552734, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.014925207942724228, + "learning_rate": 0.0006, + "loss": 6.531795501708984, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.015454350039362907, + "learning_rate": 0.0006, + "loss": 6.575699806213379, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.01623542234301567, + "learning_rate": 0.0006, + "loss": 6.582373142242432, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015452183783054352, + "learning_rate": 0.0006, + "loss": 6.576451301574707, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.021102432161569595, + "learning_rate": 0.0006, + "loss": 6.543055534362793, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.009827060624957085, + "learning_rate": 0.0006, + "loss": 6.526939392089844, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.013735136948525906, + "learning_rate": 0.0006, + "loss": 6.538825988769531, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02073509432375431, + "learning_rate": 0.0006, + "loss": 6.521315574645996, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.03110959194600582, + "learning_rate": 0.0006, + "loss": 6.561807632446289, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.045058365911245346, + "learning_rate": 0.0006, + "loss": 6.526515960693359, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.06359834969043732, + "learning_rate": 0.0006, + "loss": 6.554903030395508, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.06406767666339874, + "learning_rate": 0.0006, + "loss": 6.576272010803223, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.03403699770569801, + "learning_rate": 0.0006, + "loss": 6.5306243896484375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.036109618842601776, + "learning_rate": 0.0006, + "loss": 6.54176139831543, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.03478201478719711, + "learning_rate": 0.0006, + "loss": 6.530734062194824, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.03401544317603111, + "learning_rate": 0.0006, + "loss": 6.557794094085693, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.031629662960767746, + "learning_rate": 0.0006, + "loss": 6.564887046813965, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.02305024117231369, + "learning_rate": 0.0006, + "loss": 6.547412872314453, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030852800235152245, + "learning_rate": 0.0006, + "loss": 6.495370388031006, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.03368550166487694, + "learning_rate": 0.0006, + "loss": 6.495980262756348, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.020976359024643898, + "learning_rate": 0.0006, + "loss": 6.509410858154297, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.019652143120765686, + "learning_rate": 0.0006, + "loss": 6.493232727050781, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.024320121854543686, + "learning_rate": 0.0006, + "loss": 6.521531581878662, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.016611427068710327, + "learning_rate": 0.0006, + "loss": 6.507271766662598, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01635856367647648, + "learning_rate": 0.0006, + "loss": 6.492038726806641, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.017642095685005188, + "learning_rate": 0.0006, + "loss": 6.500349998474121, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.01679178513586521, + "learning_rate": 0.0006, + "loss": 6.500361442565918, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.011776307597756386, + "learning_rate": 0.0006, + "loss": 6.482490539550781, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.0176856629550457, + "learning_rate": 0.0006, + "loss": 6.48488712310791, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.015193155966699123, + "learning_rate": 0.0006, + "loss": 6.459545612335205, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.018623212352395058, + "learning_rate": 0.0006, + "loss": 6.478870391845703, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.012841006740927696, + "learning_rate": 0.0006, + "loss": 6.474418640136719, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.015540511347353458, + "learning_rate": 0.0006, + "loss": 6.4706525802612305, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.022695740684866905, + "learning_rate": 0.0006, + "loss": 6.455738067626953, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.026930129155516624, + "learning_rate": 0.0006, + "loss": 6.458016872406006, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.03555797412991524, + "learning_rate": 0.0006, + "loss": 6.480044364929199, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.05024682730436325, + "learning_rate": 0.0006, + "loss": 6.454564571380615, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.06616255640983582, + "learning_rate": 0.0006, + "loss": 6.433927536010742, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.059065770357847214, + "learning_rate": 0.0006, + "loss": 6.508362770080566, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.026689983904361725, + "learning_rate": 0.0006, + "loss": 6.464297294616699, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.04616398364305496, + "learning_rate": 0.0006, + "loss": 6.45448112487793, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.04980356618762016, + "learning_rate": 0.0006, + "loss": 6.4696574211120605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.0551983006298542, + "learning_rate": 0.0006, + "loss": 6.419000625610352, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.06988243758678436, + "learning_rate": 0.0006, + "loss": 6.442890167236328, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.06040226295590401, + "learning_rate": 0.0006, + "loss": 6.513258934020996, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04635024070739746, + "learning_rate": 0.0006, + "loss": 6.4710283279418945, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.04762702435255051, + "learning_rate": 0.0006, + "loss": 6.469327926635742, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.04857536032795906, + "learning_rate": 0.0006, + "loss": 6.426105499267578, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.04412693530321121, + "learning_rate": 0.0006, + "loss": 6.43179988861084, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026129672303795815, + "learning_rate": 0.0006, + "loss": 6.437918663024902, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 7.219249248504639, + "eval_runtime": 58.6837, + "eval_samples_per_second": 41.613, + "eval_steps_per_second": 1.312, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03329874947667122, + "learning_rate": 0.0006, + "loss": 6.438179969787598, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03723187372088432, + "learning_rate": 0.0006, + "loss": 6.394242286682129, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.014625726267695427, + "learning_rate": 0.0006, + "loss": 6.412118911743164, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.02917475439608097, + "learning_rate": 0.0006, + "loss": 6.453164577484131, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01811208389699459, + "learning_rate": 0.0006, + "loss": 6.396003246307373, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.027235351502895355, + "learning_rate": 0.0006, + "loss": 6.402976036071777, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.02835794910788536, + "learning_rate": 0.0006, + "loss": 6.380568504333496, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.020868154242634773, + "learning_rate": 0.0006, + "loss": 6.38240909576416, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.02113114856183529, + "learning_rate": 0.0006, + "loss": 6.377391815185547, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.018894119188189507, + "learning_rate": 0.0006, + "loss": 6.354569435119629, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.01820225641131401, + "learning_rate": 0.0006, + "loss": 6.392882347106934, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.017510728910565376, + "learning_rate": 0.0006, + "loss": 6.364162445068359, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.017893686890602112, + "learning_rate": 0.0006, + "loss": 6.373456954956055, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.013382102362811565, + "learning_rate": 0.0006, + "loss": 6.3714399337768555, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.022540833801031113, + "learning_rate": 0.0006, + "loss": 6.348170757293701, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.02110104262828827, + "learning_rate": 0.0006, + "loss": 6.413206100463867, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.014498349279165268, + "learning_rate": 0.0006, + "loss": 6.361384391784668, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.01819637045264244, + "learning_rate": 0.0006, + "loss": 6.315275192260742, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.0195972491055727, + "learning_rate": 0.0006, + "loss": 6.321994781494141, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.018822453916072845, + "learning_rate": 0.0006, + "loss": 6.345219612121582, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.019247712567448616, + "learning_rate": 0.0006, + "loss": 6.348073959350586, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.03175375238060951, + "learning_rate": 0.0006, + "loss": 6.300344944000244, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.06326444447040558, + "learning_rate": 0.0006, + "loss": 6.392579078674316, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.09667232632637024, + "learning_rate": 0.0006, + "loss": 6.36024284362793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.05774446204304695, + "learning_rate": 0.0006, + "loss": 6.365338325500488, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.08793356269598007, + "learning_rate": 0.0006, + "loss": 6.43284273147583, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.0640595331788063, + "learning_rate": 0.0006, + "loss": 6.342428207397461, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.0447838194668293, + "learning_rate": 0.0006, + "loss": 6.367971897125244, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.04768967255949974, + "learning_rate": 0.0006, + "loss": 6.35027551651001, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.027981851249933243, + "learning_rate": 0.0006, + "loss": 6.33723258972168, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.03479623422026634, + "learning_rate": 0.0006, + "loss": 6.36590051651001, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.020846687257289886, + "learning_rate": 0.0006, + "loss": 6.315143585205078, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.0312928780913353, + "learning_rate": 0.0006, + "loss": 6.340150833129883, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.020593881607055664, + "learning_rate": 0.0006, + "loss": 6.348175048828125, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.022010236978530884, + "learning_rate": 0.0006, + "loss": 6.318607330322266, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.018427059054374695, + "learning_rate": 0.0006, + "loss": 6.331822395324707, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.019964968785643578, + "learning_rate": 0.0006, + "loss": 6.287320137023926, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.02130270190536976, + "learning_rate": 0.0006, + "loss": 6.353344917297363, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.01616514101624489, + "learning_rate": 0.0006, + "loss": 6.314052581787109, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.016639474779367447, + "learning_rate": 0.0006, + "loss": 6.320351600646973, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.017620692029595375, + "learning_rate": 0.0006, + "loss": 6.311459064483643, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.013542844913899899, + "learning_rate": 0.0006, + "loss": 6.324368476867676, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01308678649365902, + "learning_rate": 0.0006, + "loss": 6.307419776916504, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.014672392047941685, + "learning_rate": 0.0006, + "loss": 6.2785420417785645, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01399585697799921, + "learning_rate": 0.0006, + "loss": 6.258619785308838, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.01330904383212328, + "learning_rate": 0.0006, + "loss": 6.299595832824707, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016613377258181572, + "learning_rate": 0.0006, + "loss": 6.2900309562683105, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.015076331794261932, + "learning_rate": 0.0006, + "loss": 6.231028079986572, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.01589728519320488, + "learning_rate": 0.0006, + "loss": 6.257922172546387, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.030972328037023544, + "learning_rate": 0.0006, + "loss": 6.285238265991211, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.05048783868551254, + "learning_rate": 0.0006, + "loss": 6.287471771240234, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.07390714436769485, + "learning_rate": 0.0006, + "loss": 6.284293174743652, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.07166943699121475, + "learning_rate": 0.0006, + "loss": 6.290704727172852, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.028792452067136765, + "learning_rate": 0.0006, + "loss": 6.243415832519531, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.04822041466832161, + "learning_rate": 0.0006, + "loss": 6.278229713439941, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.036688923835754395, + "learning_rate": 0.0006, + "loss": 6.275590896606445, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.03679114952683449, + "learning_rate": 0.0006, + "loss": 6.236509323120117, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.023249102756381035, + "learning_rate": 0.0006, + "loss": 6.230788230895996, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.02682286500930786, + "learning_rate": 0.0006, + "loss": 6.278890609741211, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.02202044241130352, + "learning_rate": 0.0006, + "loss": 6.234531402587891, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.02545459009706974, + "learning_rate": 0.0006, + "loss": 6.271790981292725, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.02012932486832142, + "learning_rate": 0.0006, + "loss": 6.247462749481201, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.02216474711894989, + "learning_rate": 0.0006, + "loss": 6.26396369934082, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.016078704968094826, + "learning_rate": 0.0006, + "loss": 6.228609085083008, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.023021478205919266, + "learning_rate": 0.0006, + "loss": 6.197244167327881, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.024766333401203156, + "learning_rate": 0.0006, + "loss": 6.181090354919434, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.025581466034054756, + "learning_rate": 0.0006, + "loss": 6.201817512512207, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.016964981332421303, + "learning_rate": 0.0006, + "loss": 6.201486587524414, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.023957721889019012, + "learning_rate": 0.0006, + "loss": 6.185701847076416, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.029775245115160942, + "learning_rate": 0.0006, + "loss": 6.242060661315918, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04007191210985184, + "learning_rate": 0.0006, + "loss": 6.19931697845459, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04882276430726051, + "learning_rate": 0.0006, + "loss": 6.1880598068237305, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 7.018043518066406, + "eval_runtime": 59.8815, + "eval_samples_per_second": 40.781, + "eval_steps_per_second": 1.286, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.07144106924533844, + "learning_rate": 0.0006, + "loss": 6.172799110412598, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.06712841987609863, + "learning_rate": 0.0006, + "loss": 6.228682518005371, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.04405239224433899, + "learning_rate": 0.0006, + "loss": 6.1827898025512695, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.03418995440006256, + "learning_rate": 0.0006, + "loss": 6.200876235961914, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.04002382606267929, + "learning_rate": 0.0006, + "loss": 6.1799211502075195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.029748762026429176, + "learning_rate": 0.0006, + "loss": 6.125422477722168, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03408357501029968, + "learning_rate": 0.0006, + "loss": 6.203458309173584, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.02227279357612133, + "learning_rate": 0.0006, + "loss": 6.135439872741699, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02727862447500229, + "learning_rate": 0.0006, + "loss": 6.202380180358887, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.026019075885415077, + "learning_rate": 0.0006, + "loss": 6.201343536376953, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.027227282524108887, + "learning_rate": 0.0006, + "loss": 6.164059638977051, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.027649980038404465, + "learning_rate": 0.0006, + "loss": 6.135448932647705, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02168039232492447, + "learning_rate": 0.0006, + "loss": 6.154831886291504, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02101467363536358, + "learning_rate": 0.0006, + "loss": 6.128695487976074, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.018344245851039886, + "learning_rate": 0.0006, + "loss": 6.136524200439453, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020323453471064568, + "learning_rate": 0.0006, + "loss": 6.1206488609313965, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.01750849559903145, + "learning_rate": 0.0006, + "loss": 6.128170967102051, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017778221517801285, + "learning_rate": 0.0006, + "loss": 6.150458812713623, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.01863938942551613, + "learning_rate": 0.0006, + "loss": 6.100106716156006, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.020196080207824707, + "learning_rate": 0.0006, + "loss": 6.149346351623535, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.03203890845179558, + "learning_rate": 0.0006, + "loss": 6.124648094177246, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.05325170233845711, + "learning_rate": 0.0006, + "loss": 6.115618705749512, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.05888345465064049, + "learning_rate": 0.0006, + "loss": 6.165221214294434, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.026873115450143814, + "learning_rate": 0.0006, + "loss": 6.039590835571289, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.06304649263620377, + "learning_rate": 0.0006, + "loss": 6.131900787353516, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.08567557483911514, + "learning_rate": 0.0006, + "loss": 6.218114852905273, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.08868218213319778, + "learning_rate": 0.0006, + "loss": 6.216706275939941, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.03955601155757904, + "learning_rate": 0.0006, + "loss": 6.1566925048828125, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.04979498311877251, + "learning_rate": 0.0006, + "loss": 6.19327449798584, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.03253026679158211, + "learning_rate": 0.0006, + "loss": 6.133993148803711, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03016652725636959, + "learning_rate": 0.0006, + "loss": 6.174962997436523, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.03159862384200096, + "learning_rate": 0.0006, + "loss": 6.082124710083008, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.027673959732055664, + "learning_rate": 0.0006, + "loss": 6.111873149871826, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.022789714857935905, + "learning_rate": 0.0006, + "loss": 6.129437446594238, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.030061012133955956, + "learning_rate": 0.0006, + "loss": 6.126370429992676, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.019007068127393723, + "learning_rate": 0.0006, + "loss": 6.07887077331543, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.01897653006017208, + "learning_rate": 0.0006, + "loss": 6.096936225891113, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.024659698829054832, + "learning_rate": 0.0006, + "loss": 6.0933380126953125, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.021924549713730812, + "learning_rate": 0.0006, + "loss": 6.105311870574951, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.018817558884620667, + "learning_rate": 0.0006, + "loss": 6.058895111083984, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.02477840520441532, + "learning_rate": 0.0006, + "loss": 6.038928031921387, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.018968632444739342, + "learning_rate": 0.0006, + "loss": 6.1368937492370605, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.02409801445901394, + "learning_rate": 0.0006, + "loss": 6.110324859619141, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.017595326527953148, + "learning_rate": 0.0006, + "loss": 6.074368476867676, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.01875406503677368, + "learning_rate": 0.0006, + "loss": 6.121023654937744, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.018781628459692, + "learning_rate": 0.0006, + "loss": 6.064914703369141, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.017003126442432404, + "learning_rate": 0.0006, + "loss": 6.039608001708984, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.017665037885308266, + "learning_rate": 0.0006, + "loss": 6.053671836853027, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.02066265605390072, + "learning_rate": 0.0006, + "loss": 6.090536117553711, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02888769656419754, + "learning_rate": 0.0006, + "loss": 5.9858903884887695, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.03744485229253769, + "learning_rate": 0.0006, + "loss": 6.086264610290527, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04719431325793266, + "learning_rate": 0.0006, + "loss": 6.09747314453125, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.044963907450437546, + "learning_rate": 0.0006, + "loss": 6.091272354125977, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03810194507241249, + "learning_rate": 0.0006, + "loss": 6.084904193878174, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.04626769199967384, + "learning_rate": 0.0006, + "loss": 6.017384052276611, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.03459552302956581, + "learning_rate": 0.0006, + "loss": 6.111314296722412, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.027385666966438293, + "learning_rate": 0.0006, + "loss": 6.002509117126465, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.03417652100324631, + "learning_rate": 0.0006, + "loss": 6.0421061515808105, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.04415450245141983, + "learning_rate": 0.0006, + "loss": 6.035305023193359, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.03259635716676712, + "learning_rate": 0.0006, + "loss": 6.025840759277344, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.024941129609942436, + "learning_rate": 0.0006, + "loss": 6.043416976928711, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.02434500865638256, + "learning_rate": 0.0006, + "loss": 6.002052307128906, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.01876584254205227, + "learning_rate": 0.0006, + "loss": 6.045035362243652, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.02254367619752884, + "learning_rate": 0.0006, + "loss": 5.972503662109375, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.024504603818058968, + "learning_rate": 0.0006, + "loss": 5.992846488952637, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.02317827194929123, + "learning_rate": 0.0006, + "loss": 6.007697582244873, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02225821278989315, + "learning_rate": 0.0006, + "loss": 6.01463508605957, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.01811717264354229, + "learning_rate": 0.0006, + "loss": 5.995022773742676, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.019589431583881378, + "learning_rate": 0.0006, + "loss": 5.990633964538574, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015422016382217407, + "learning_rate": 0.0006, + "loss": 6.011263847351074, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.9954938888549805, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.022172044962644577, + "learning_rate": 0.0006, + "loss": 5.95679235458374, + "step": 576 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.4556906890657792e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-576/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef +size 4792 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/config.json new file mode 100644 index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151674 +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b853dbdc80b4273e5d550f1e2dbd877a098bb903 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e34000ad5b58ca0564625863e22c69e39127b37d86835e20118d4fd08824445 +size 583362376 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..6f03aa4ff10b813a3e1c00c87d636b2965e1eec4 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed7b28dab7d301258603fe180514d85e1b9a3b9f82f0a8c925caeda756dc10a7 +size 1166837626 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..35b54c5c64a65f3977bb4b4ec2b748cd17b20f52 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e5d12f33f319393c8dafffd6b14ce4cfa3aeb46c8c3760b273eddbb8152d1c7 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..aa9143a4e362d6c0ee326f19cdd26dcecaf7e6a3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce7ffd7ac75644c8f6fc42f002287b7b0c3c585e58668b8936b9fddaed2f8cc5 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..d65c7a733824cbda65cbe53f302ffcca01bce8c6 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc504e473f67f378865973402a0109aa8223a3b343d77a078ed4b4bd0df27098 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..476c1ccbff619bfca532c36bdd72e59d0543bfd6 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afa0376616f56e7a706215e7b0a97a0be812db73d881b1ab615ea8260bc9350f +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f +size 1064 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14 +size 11424648 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/tokenizer_config.json @@ -0,0 +1,21 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>", + "<|mask|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..34f68e1a9ca8b8c4c94fbde71d1947f973c6be4c --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/trainer_state.json @@ -0,0 +1,538 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 72, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.5768774151802063, + "learning_rate": 0.0, + "loss": 12.050328254699707, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.5784570574760437, + "learning_rate": 5.999999999999999e-06, + "loss": 12.049251556396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.5607187151908875, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.921792030334473, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.46466919779777527, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.686628341674805, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.38012945652008057, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.432823181152344, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.33215776085853577, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.207542419433594, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.2928776741027832, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.000097274780273, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.26018962264060974, + "learning_rate": 4.2e-05, + "loss": 10.825748443603516, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.23717083036899567, + "learning_rate": 4.7999999999999994e-05, + "loss": 10.677122116088867, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.21641355752944946, + "learning_rate": 5.399999999999999e-05, + "loss": 10.569148063659668, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.1982956826686859, + "learning_rate": 5.9999999999999995e-05, + "loss": 10.49653434753418, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.18819282948970795, + "learning_rate": 6.599999999999999e-05, + "loss": 10.432951927185059, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.18322618305683136, + "learning_rate": 7.199999999999999e-05, + "loss": 10.387115478515625, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.18399226665496826, + "learning_rate": 7.8e-05, + "loss": 10.33381462097168, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.1833190619945526, + "learning_rate": 8.4e-05, + "loss": 10.301689147949219, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.18490834534168243, + "learning_rate": 8.999999999999999e-05, + "loss": 10.240968704223633, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.18358023464679718, + "learning_rate": 9.599999999999999e-05, + "loss": 10.194717407226562, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.1801682710647583, + "learning_rate": 0.000102, + "loss": 10.14897632598877, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.17899440228939056, + "learning_rate": 0.00010799999999999998, + "loss": 10.078763961791992, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.17666497826576233, + "learning_rate": 0.00011399999999999999, + "loss": 10.01095962524414, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.17196154594421387, + "learning_rate": 0.00011999999999999999, + "loss": 9.951203346252441, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.169203519821167, + "learning_rate": 0.00012599999999999997, + "loss": 9.871726036071777, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.16500306129455566, + "learning_rate": 0.00013199999999999998, + "loss": 9.803646087646484, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.16128124296665192, + "learning_rate": 0.000138, + "loss": 9.728487014770508, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.15904058516025543, + "learning_rate": 0.00014399999999999998, + "loss": 9.642473220825195, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.15560561418533325, + "learning_rate": 0.00015, + "loss": 9.56414794921875, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.15278612077236176, + "learning_rate": 0.000156, + "loss": 9.484230041503906, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.15263237059116364, + "learning_rate": 0.000162, + "loss": 9.383644104003906, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.14945024251937866, + "learning_rate": 0.000168, + "loss": 9.29840087890625, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.14614646136760712, + "learning_rate": 0.00017399999999999997, + "loss": 9.208931922912598, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.14295098185539246, + "learning_rate": 0.00017999999999999998, + "loss": 9.122323989868164, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.13868345320224762, + "learning_rate": 0.000186, + "loss": 9.03034496307373, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.135064035654068, + "learning_rate": 0.00019199999999999998, + "loss": 8.92818832397461, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.13097241520881653, + "learning_rate": 0.000198, + "loss": 8.834627151489258, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.12478776276111603, + "learning_rate": 0.000204, + "loss": 8.755172729492188, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.1214982345700264, + "learning_rate": 0.00020999999999999998, + "loss": 8.647891998291016, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.11958328634500504, + "learning_rate": 0.00021599999999999996, + "loss": 8.543274879455566, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.11680915951728821, + "learning_rate": 0.00022199999999999998, + "loss": 8.44540023803711, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.11491474509239197, + "learning_rate": 0.00022799999999999999, + "loss": 8.349921226501465, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.11160755902528763, + "learning_rate": 0.000234, + "loss": 8.26960563659668, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.10840637236833572, + "learning_rate": 0.00023999999999999998, + "loss": 8.192136764526367, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.10585298389196396, + "learning_rate": 0.00024599999999999996, + "loss": 8.09546947479248, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.1032542735338211, + "learning_rate": 0.00025199999999999995, + "loss": 7.991090774536133, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.0986943170428276, + "learning_rate": 0.000258, + "loss": 7.92384147644043, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.09320762753486633, + "learning_rate": 0.00026399999999999997, + "loss": 7.8478193283081055, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.09168332070112228, + "learning_rate": 0.00027, + "loss": 7.76185417175293, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.08832395076751709, + "learning_rate": 0.000276, + "loss": 7.679399490356445, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.08816400915384293, + "learning_rate": 0.00028199999999999997, + "loss": 7.5789642333984375, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0861717164516449, + "learning_rate": 0.00028799999999999995, + "loss": 7.519214630126953, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.08103746175765991, + "learning_rate": 0.000294, + "loss": 7.4475321769714355, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.07545104622840881, + "learning_rate": 0.0003, + "loss": 7.401182174682617, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.07042575627565384, + "learning_rate": 0.00030599999999999996, + "loss": 7.340173244476318, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.06737572699785233, + "learning_rate": 0.000312, + "loss": 7.265697002410889, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.06151146441698074, + "learning_rate": 0.000318, + "loss": 7.231061935424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.05638071522116661, + "learning_rate": 0.000324, + "loss": 7.195466995239258, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.05084463208913803, + "learning_rate": 0.00033, + "loss": 7.137603759765625, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.04293261095881462, + "learning_rate": 0.000336, + "loss": 7.115458965301514, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.036064863204956055, + "learning_rate": 0.00034199999999999996, + "loss": 7.089038848876953, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.029026754200458527, + "learning_rate": 0.00034799999999999995, + "loss": 7.0584306716918945, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.02286517433822155, + "learning_rate": 0.00035399999999999993, + "loss": 7.049450874328613, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.018878433853387833, + "learning_rate": 0.00035999999999999997, + "loss": 7.008824348449707, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.010773813351988792, + "learning_rate": 0.00036599999999999995, + "loss": 7.024829864501953, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.008794732391834259, + "learning_rate": 0.000372, + "loss": 7.018399238586426, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.011709796264767647, + "learning_rate": 0.00037799999999999997, + "loss": 7.0101752281188965, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.012123258784413338, + "learning_rate": 0.00038399999999999996, + "loss": 7.024317741394043, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01530320756137371, + "learning_rate": 0.00039, + "loss": 7.027738571166992, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.01890002004802227, + "learning_rate": 0.000396, + "loss": 7.024099349975586, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.020684150978922844, + "learning_rate": 0.000402, + "loss": 7.007943153381348, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.019826196134090424, + "learning_rate": 0.000408, + "loss": 7.006874084472656, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.019218673929572105, + "learning_rate": 0.0004139999999999999, + "loss": 7.002897262573242, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.013856411911547184, + "learning_rate": 0.00041999999999999996, + "loss": 6.993999481201172, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.010390003211796284, + "learning_rate": 0.00042599999999999995, + "loss": 7.000683784484863, + "step": 72 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.069613361332224e+16, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-72/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef +size 4792 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/config.json new file mode 100644 index 0000000000000000000000000000000000000000..175ee3431793018f23f74824f69554b21afcbd4f --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151674 +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/generation_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/model.safetensors b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c4db0d8d056f67e0168a10695de929a8ce130a69 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a15b1846227dee8f858b94836547f08520546232e3813c085490d2ba5c6d725 +size 583362376 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e8f771496f9d9ade0a40bf0d852c47e2b4551c06 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1659a40ad5174425855429c58a7e46d8cc0ef902b4bbf178f1a54296d8b98fbb +size 1166837626 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..870ed7cdb8744b07d1ff1f9112387c96a63c2807 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dad714e377efb4a1838314db7e924ef1cd99f199b53380b22d6c44b28966149b +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..5925e0f1a7b68861f2b2cc205a9eff31aa6648fb --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0999d19c49c17c3519518d23620eff41a1653e67643d57a5770816881df7313 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth new file mode 100644 index 0000000000000000000000000000000000000000..8fd148eaedcb574a10f491ff8aadebd0bcdaabbe --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_2.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:016c5f9e311d3225842840635607ff7f8e43814071a963a3c3f1559b8937b7c9 +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth new file mode 100644 index 0000000000000000000000000000000000000000..94829b15fcfd51a5cc4108f3610ac808956477b6 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/rng_state_3.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5539b9e658812aa3870e38193d2b0bdc5b6664af0e769e6a604f2a69a04a9c9d +size 15024 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f21fc027a37dccb66b77a4752de227c557d8ff04 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1a1dda88112968a1342ac518f362b4f5e138437f0b670c818ad8493b91ca49d +size 1064 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..9ea28d9eb325c36bcab4c1a08fc93fd598929f08 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60e99075d371ecbb9f1dde97363f612d45bf0bf88f1a2006d9bfcf1939333c14 +size 11424648 diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cadd997fc26040d23ad6677fcb39d8d18b01f55e --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/tokenizer_config.json @@ -0,0 +1,21 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>", + "<|next_1_pred|>", + "<|next_2_pred|>", + "<|mask|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..fb8c37d335462658260ed45e6615de9045152ec1 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/trainer_state.json @@ -0,0 +1,6170 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 12.0, + "eval_steps": 500, + "global_step": 864, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.5768774151802063, + "learning_rate": 0.0, + "loss": 12.050328254699707, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.5784570574760437, + "learning_rate": 5.999999999999999e-06, + "loss": 12.049251556396484, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.5607187151908875, + "learning_rate": 1.1999999999999999e-05, + "loss": 11.921792030334473, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.46466919779777527, + "learning_rate": 1.7999999999999997e-05, + "loss": 11.686628341674805, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.38012945652008057, + "learning_rate": 2.3999999999999997e-05, + "loss": 11.432823181152344, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.33215776085853577, + "learning_rate": 2.9999999999999997e-05, + "loss": 11.207542419433594, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.2928776741027832, + "learning_rate": 3.5999999999999994e-05, + "loss": 11.000097274780273, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.26018962264060974, + "learning_rate": 4.2e-05, + "loss": 10.825748443603516, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.23717083036899567, + "learning_rate": 4.7999999999999994e-05, + "loss": 10.677122116088867, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.21641355752944946, + "learning_rate": 5.399999999999999e-05, + "loss": 10.569148063659668, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.1982956826686859, + "learning_rate": 5.9999999999999995e-05, + "loss": 10.49653434753418, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.18819282948970795, + "learning_rate": 6.599999999999999e-05, + "loss": 10.432951927185059, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.18322618305683136, + "learning_rate": 7.199999999999999e-05, + "loss": 10.387115478515625, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.18399226665496826, + "learning_rate": 7.8e-05, + "loss": 10.33381462097168, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.1833190619945526, + "learning_rate": 8.4e-05, + "loss": 10.301689147949219, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.18490834534168243, + "learning_rate": 8.999999999999999e-05, + "loss": 10.240968704223633, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.18358023464679718, + "learning_rate": 9.599999999999999e-05, + "loss": 10.194717407226562, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.1801682710647583, + "learning_rate": 0.000102, + "loss": 10.14897632598877, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.17899440228939056, + "learning_rate": 0.00010799999999999998, + "loss": 10.078763961791992, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.17666497826576233, + "learning_rate": 0.00011399999999999999, + "loss": 10.01095962524414, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.17196154594421387, + "learning_rate": 0.00011999999999999999, + "loss": 9.951203346252441, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.169203519821167, + "learning_rate": 0.00012599999999999997, + "loss": 9.871726036071777, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.16500306129455566, + "learning_rate": 0.00013199999999999998, + "loss": 9.803646087646484, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.16128124296665192, + "learning_rate": 0.000138, + "loss": 9.728487014770508, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.15904058516025543, + "learning_rate": 0.00014399999999999998, + "loss": 9.642473220825195, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.15560561418533325, + "learning_rate": 0.00015, + "loss": 9.56414794921875, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.15278612077236176, + "learning_rate": 0.000156, + "loss": 9.484230041503906, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.15263237059116364, + "learning_rate": 0.000162, + "loss": 9.383644104003906, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.14945024251937866, + "learning_rate": 0.000168, + "loss": 9.29840087890625, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.14614646136760712, + "learning_rate": 0.00017399999999999997, + "loss": 9.208931922912598, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.14295098185539246, + "learning_rate": 0.00017999999999999998, + "loss": 9.122323989868164, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.13868345320224762, + "learning_rate": 0.000186, + "loss": 9.03034496307373, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.135064035654068, + "learning_rate": 0.00019199999999999998, + "loss": 8.92818832397461, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.13097241520881653, + "learning_rate": 0.000198, + "loss": 8.834627151489258, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.12478776276111603, + "learning_rate": 0.000204, + "loss": 8.755172729492188, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.1214982345700264, + "learning_rate": 0.00020999999999999998, + "loss": 8.647891998291016, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.11958328634500504, + "learning_rate": 0.00021599999999999996, + "loss": 8.543274879455566, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.11680915951728821, + "learning_rate": 0.00022199999999999998, + "loss": 8.44540023803711, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.11491474509239197, + "learning_rate": 0.00022799999999999999, + "loss": 8.349921226501465, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.11160755902528763, + "learning_rate": 0.000234, + "loss": 8.26960563659668, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.10840637236833572, + "learning_rate": 0.00023999999999999998, + "loss": 8.192136764526367, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.10585298389196396, + "learning_rate": 0.00024599999999999996, + "loss": 8.09546947479248, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.1032542735338211, + "learning_rate": 0.00025199999999999995, + "loss": 7.991090774536133, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.0986943170428276, + "learning_rate": 0.000258, + "loss": 7.92384147644043, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.09320762753486633, + "learning_rate": 0.00026399999999999997, + "loss": 7.8478193283081055, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.09168332070112228, + "learning_rate": 0.00027, + "loss": 7.76185417175293, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.08832395076751709, + "learning_rate": 0.000276, + "loss": 7.679399490356445, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.08816400915384293, + "learning_rate": 0.00028199999999999997, + "loss": 7.5789642333984375, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.0861717164516449, + "learning_rate": 0.00028799999999999995, + "loss": 7.519214630126953, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.08103746175765991, + "learning_rate": 0.000294, + "loss": 7.4475321769714355, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.07545104622840881, + "learning_rate": 0.0003, + "loss": 7.401182174682617, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.07042575627565384, + "learning_rate": 0.00030599999999999996, + "loss": 7.340173244476318, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.06737572699785233, + "learning_rate": 0.000312, + "loss": 7.265697002410889, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.06151146441698074, + "learning_rate": 0.000318, + "loss": 7.231061935424805, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.05638071522116661, + "learning_rate": 0.000324, + "loss": 7.195466995239258, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.05084463208913803, + "learning_rate": 0.00033, + "loss": 7.137603759765625, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.04293261095881462, + "learning_rate": 0.000336, + "loss": 7.115458965301514, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.036064863204956055, + "learning_rate": 0.00034199999999999996, + "loss": 7.089038848876953, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.029026754200458527, + "learning_rate": 0.00034799999999999995, + "loss": 7.0584306716918945, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.02286517433822155, + "learning_rate": 0.00035399999999999993, + "loss": 7.049450874328613, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.018878433853387833, + "learning_rate": 0.00035999999999999997, + "loss": 7.008824348449707, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.010773813351988792, + "learning_rate": 0.00036599999999999995, + "loss": 7.024829864501953, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.008794732391834259, + "learning_rate": 0.000372, + "loss": 7.018399238586426, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.011709796264767647, + "learning_rate": 0.00037799999999999997, + "loss": 7.0101752281188965, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.012123258784413338, + "learning_rate": 0.00038399999999999996, + "loss": 7.024317741394043, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.01530320756137371, + "learning_rate": 0.00039, + "loss": 7.027738571166992, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.01890002004802227, + "learning_rate": 0.000396, + "loss": 7.024099349975586, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.020684150978922844, + "learning_rate": 0.000402, + "loss": 7.007943153381348, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.019826196134090424, + "learning_rate": 0.000408, + "loss": 7.006874084472656, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.019218673929572105, + "learning_rate": 0.0004139999999999999, + "loss": 7.002897262573242, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.013856411911547184, + "learning_rate": 0.00041999999999999996, + "loss": 6.993999481201172, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.010390003211796284, + "learning_rate": 0.00042599999999999995, + "loss": 7.000683784484863, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.91140079498291, + "eval_runtime": 57.7925, + "eval_samples_per_second": 42.255, + "eval_steps_per_second": 1.332, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.013802163302898407, + "learning_rate": 0.00043199999999999993, + "loss": 6.998257637023926, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.010109687224030495, + "learning_rate": 0.00043799999999999997, + "loss": 7.002542018890381, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.008019139990210533, + "learning_rate": 0.00044399999999999995, + "loss": 6.985941410064697, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.0069672465324401855, + "learning_rate": 0.00045, + "loss": 6.99245023727417, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.00714076729491353, + "learning_rate": 0.00045599999999999997, + "loss": 6.959296226501465, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.008408435620367527, + "learning_rate": 0.00046199999999999995, + "loss": 6.954504489898682, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.008778786286711693, + "learning_rate": 0.000468, + "loss": 7.0007004737854, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.007753407116979361, + "learning_rate": 0.000474, + "loss": 6.98183536529541, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.01087260153144598, + "learning_rate": 0.00047999999999999996, + "loss": 6.991066932678223, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.006122548598796129, + "learning_rate": 0.000486, + "loss": 6.978824138641357, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.0067383465357124805, + "learning_rate": 0.0004919999999999999, + "loss": 6.961703777313232, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.007922791875898838, + "learning_rate": 0.000498, + "loss": 6.954860687255859, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.00818922370672226, + "learning_rate": 0.0005039999999999999, + "loss": 6.945961952209473, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.006791461259126663, + "learning_rate": 0.0005099999999999999, + "loss": 6.968680381774902, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.005394163075834513, + "learning_rate": 0.000516, + "loss": 6.956404685974121, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.004500220064073801, + "learning_rate": 0.000522, + "loss": 6.96998405456543, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008699574507772923, + "learning_rate": 0.0005279999999999999, + "loss": 6.955585479736328, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.005889365915209055, + "learning_rate": 0.000534, + "loss": 6.9708051681518555, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.009646810591220856, + "learning_rate": 0.00054, + "loss": 6.962608814239502, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.004517171997576952, + "learning_rate": 0.0005459999999999999, + "loss": 6.957627296447754, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.006339677143841982, + "learning_rate": 0.000552, + "loss": 6.963141441345215, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.005387383047491312, + "learning_rate": 0.000558, + "loss": 6.934465408325195, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.005252422299236059, + "learning_rate": 0.0005639999999999999, + "loss": 6.956108093261719, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.008699414320290089, + "learning_rate": 0.00057, + "loss": 6.973369598388672, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.003947409335523844, + "learning_rate": 0.0005759999999999999, + "loss": 6.950225830078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.00768681475892663, + "learning_rate": 0.0005819999999999999, + "loss": 6.964739799499512, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.006563829258084297, + "learning_rate": 0.000588, + "loss": 6.972421169281006, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.007606555242091417, + "learning_rate": 0.0005939999999999999, + "loss": 6.958213806152344, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.006250880658626556, + "learning_rate": 0.0006, + "loss": 6.962296485900879, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.0073033408261835575, + "learning_rate": 0.0006, + "loss": 6.969705104827881, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.0054510910995304585, + "learning_rate": 0.0006, + "loss": 6.9652862548828125, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.004895673133432865, + "learning_rate": 0.0006, + "loss": 6.956023693084717, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.005306515377014875, + "learning_rate": 0.0006, + "loss": 6.963231086730957, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.004742146469652653, + "learning_rate": 0.0006, + "loss": 6.9582061767578125, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.004220000468194485, + "learning_rate": 0.0006, + "loss": 6.972961902618408, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.0042153168469667435, + "learning_rate": 0.0006, + "loss": 6.961165904998779, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.005165040958672762, + "learning_rate": 0.0006, + "loss": 6.95018196105957, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.003930266480892897, + "learning_rate": 0.0006, + "loss": 6.9651689529418945, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.004465071018785238, + "learning_rate": 0.0006, + "loss": 6.960553169250488, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.004888612311333418, + "learning_rate": 0.0006, + "loss": 6.967258930206299, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.007497863378375769, + "learning_rate": 0.0006, + "loss": 6.954545021057129, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.005659648682922125, + "learning_rate": 0.0006, + "loss": 6.962299823760986, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.006641396787017584, + "learning_rate": 0.0006, + "loss": 6.952983856201172, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.009715928696095943, + "learning_rate": 0.0006, + "loss": 6.958622932434082, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.012789523229002953, + "learning_rate": 0.0006, + "loss": 6.932043075561523, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.021673237904906273, + "learning_rate": 0.0006, + "loss": 6.931290149688721, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.16024991869926453, + "learning_rate": 0.0006, + "loss": 6.979497909545898, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.2434523105621338, + "learning_rate": 0.0006, + "loss": 7.185218811035156, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.39252421259880066, + "learning_rate": 0.0006, + "loss": 7.137226581573486, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.10565312206745148, + "learning_rate": 0.0006, + "loss": 6.984494209289551, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.1033431813120842, + "learning_rate": 0.0006, + "loss": 7.001287937164307, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.14875490963459015, + "learning_rate": 0.0006, + "loss": 7.051436901092529, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.1374562531709671, + "learning_rate": 0.0006, + "loss": 7.0293121337890625, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.0910211056470871, + "learning_rate": 0.0006, + "loss": 7.007458686828613, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.021633487194776535, + "learning_rate": 0.0006, + "loss": 6.996816635131836, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.08692754060029984, + "learning_rate": 0.0006, + "loss": 6.998541831970215, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.12613292038440704, + "learning_rate": 0.0006, + "loss": 6.9970703125, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.09462810307741165, + "learning_rate": 0.0006, + "loss": 6.997045040130615, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.03534962609410286, + "learning_rate": 0.0006, + "loss": 6.9738450050354, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.033839885145425797, + "learning_rate": 0.0006, + "loss": 6.946001052856445, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.06263269484043121, + "learning_rate": 0.0006, + "loss": 6.968981742858887, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.07117997109889984, + "learning_rate": 0.0006, + "loss": 6.981535911560059, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.06336277723312378, + "learning_rate": 0.0006, + "loss": 6.976946830749512, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.047467734664678574, + "learning_rate": 0.0006, + "loss": 6.961673736572266, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.018787868320941925, + "learning_rate": 0.0006, + "loss": 6.948919296264648, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.025909408926963806, + "learning_rate": 0.0006, + "loss": 6.976190567016602, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.052842333912849426, + "learning_rate": 0.0006, + "loss": 6.971207618713379, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.06421888619661331, + "learning_rate": 0.0006, + "loss": 6.979109764099121, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.055596910417079926, + "learning_rate": 0.0006, + "loss": 6.972118854522705, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.03267091140151024, + "learning_rate": 0.0006, + "loss": 6.947260856628418, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.008032613433897495, + "learning_rate": 0.0006, + "loss": 6.936816215515137, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.020184025168418884, + "learning_rate": 0.0006, + "loss": 6.929848670959473, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 7.836523056030273, + "eval_runtime": 58.5622, + "eval_samples_per_second": 41.699, + "eval_steps_per_second": 1.315, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.031980376690626144, + "learning_rate": 0.0006, + "loss": 6.968081474304199, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.03802533075213432, + "learning_rate": 0.0006, + "loss": 6.95728874206543, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.03796226158738136, + "learning_rate": 0.0006, + "loss": 6.96516752243042, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.02621951326727867, + "learning_rate": 0.0006, + "loss": 6.956058502197266, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.013292718678712845, + "learning_rate": 0.0006, + "loss": 6.9783935546875, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.009446968324482441, + "learning_rate": 0.0006, + "loss": 6.96873664855957, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.021703092381358147, + "learning_rate": 0.0006, + "loss": 6.959234237670898, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.027765290811657906, + "learning_rate": 0.0006, + "loss": 6.968524932861328, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.028761165216565132, + "learning_rate": 0.0006, + "loss": 6.954620838165283, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.019729383289813995, + "learning_rate": 0.0006, + "loss": 6.949009895324707, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.008990089409053326, + "learning_rate": 0.0006, + "loss": 6.956226348876953, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.00549644511193037, + "learning_rate": 0.0006, + "loss": 6.957143783569336, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.014436670579016209, + "learning_rate": 0.0006, + "loss": 6.954257011413574, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.01990494504570961, + "learning_rate": 0.0006, + "loss": 6.942131519317627, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.019155776128172874, + "learning_rate": 0.0006, + "loss": 6.946969985961914, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.014970783144235611, + "learning_rate": 0.0006, + "loss": 6.957938194274902, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.009875659830868244, + "learning_rate": 0.0006, + "loss": 6.953309059143066, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.004792087245732546, + "learning_rate": 0.0006, + "loss": 6.975146770477295, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.009936527349054813, + "learning_rate": 0.0006, + "loss": 6.961674690246582, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.015100127086043358, + "learning_rate": 0.0006, + "loss": 6.968759536743164, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.01449542585760355, + "learning_rate": 0.0006, + "loss": 6.966510772705078, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.01073537115007639, + "learning_rate": 0.0006, + "loss": 6.936373710632324, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.006766272243112326, + "learning_rate": 0.0006, + "loss": 6.96397590637207, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.005055439658463001, + "learning_rate": 0.0006, + "loss": 6.961050987243652, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.006252622697502375, + "learning_rate": 0.0006, + "loss": 6.980506420135498, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.008084346540272236, + "learning_rate": 0.0006, + "loss": 6.9346604347229, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.010260380804538727, + "learning_rate": 0.0006, + "loss": 6.964118957519531, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.010697875171899796, + "learning_rate": 0.0006, + "loss": 6.941572189331055, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.007338830269873142, + "learning_rate": 0.0006, + "loss": 6.950471878051758, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.005931832827627659, + "learning_rate": 0.0006, + "loss": 6.962935924530029, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.0061243255622684956, + "learning_rate": 0.0006, + "loss": 6.946418285369873, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.007983328774571419, + "learning_rate": 0.0006, + "loss": 6.959758758544922, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.010470672510564327, + "learning_rate": 0.0006, + "loss": 6.93851375579834, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.011147388257086277, + "learning_rate": 0.0006, + "loss": 6.93258810043335, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01058544497936964, + "learning_rate": 0.0006, + "loss": 6.932259559631348, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.016850745305418968, + "learning_rate": 0.0006, + "loss": 6.924542427062988, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.023101547732949257, + "learning_rate": 0.0006, + "loss": 6.914878845214844, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.015795735642313957, + "learning_rate": 0.0006, + "loss": 6.922432899475098, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.008821426890790462, + "learning_rate": 0.0006, + "loss": 6.924562454223633, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.012502548284828663, + "learning_rate": 0.0006, + "loss": 6.914718151092529, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.011545066721737385, + "learning_rate": 0.0006, + "loss": 6.915705680847168, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.022190893068909645, + "learning_rate": 0.0006, + "loss": 6.88851261138916, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.023395851254463196, + "learning_rate": 0.0006, + "loss": 6.877655029296875, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015601156279444695, + "learning_rate": 0.0006, + "loss": 6.889644145965576, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.01362934336066246, + "learning_rate": 0.0006, + "loss": 6.891547203063965, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.012305490672588348, + "learning_rate": 0.0006, + "loss": 6.890357971191406, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.01623358018696308, + "learning_rate": 0.0006, + "loss": 6.849443435668945, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.018063291907310486, + "learning_rate": 0.0006, + "loss": 6.842372894287109, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.010179434902966022, + "learning_rate": 0.0006, + "loss": 6.870055198669434, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.009986722841858864, + "learning_rate": 0.0006, + "loss": 6.859061241149902, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.01117654051631689, + "learning_rate": 0.0006, + "loss": 6.861199378967285, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.008551654405891895, + "learning_rate": 0.0006, + "loss": 6.850255012512207, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.018904326483607292, + "learning_rate": 0.0006, + "loss": 6.848639965057373, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.02080724574625492, + "learning_rate": 0.0006, + "loss": 6.80456018447876, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.012010873295366764, + "learning_rate": 0.0006, + "loss": 6.848775863647461, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.008113056421279907, + "learning_rate": 0.0006, + "loss": 6.849387168884277, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.012912592850625515, + "learning_rate": 0.0006, + "loss": 6.832240104675293, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.01146865077316761, + "learning_rate": 0.0006, + "loss": 6.850383281707764, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.0171892661601305, + "learning_rate": 0.0006, + "loss": 6.839577674865723, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.012083170004189014, + "learning_rate": 0.0006, + "loss": 6.818614482879639, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.0062555656768381596, + "learning_rate": 0.0006, + "loss": 6.831291198730469, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.010102466680109501, + "learning_rate": 0.0006, + "loss": 6.846522808074951, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.009042087942361832, + "learning_rate": 0.0006, + "loss": 6.822072982788086, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.011133556254208088, + "learning_rate": 0.0006, + "loss": 6.811252593994141, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.00990577507764101, + "learning_rate": 0.0006, + "loss": 6.814435958862305, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.006605073343962431, + "learning_rate": 0.0006, + "loss": 6.831300735473633, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.007831175811588764, + "learning_rate": 0.0006, + "loss": 6.841738700866699, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.005536255426704884, + "learning_rate": 0.0006, + "loss": 6.8043670654296875, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.006532667670398951, + "learning_rate": 0.0006, + "loss": 6.8091840744018555, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.006674565840512514, + "learning_rate": 0.0006, + "loss": 6.811327934265137, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.01150344219058752, + "learning_rate": 0.0006, + "loss": 6.810762405395508, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.009015528485178947, + "learning_rate": 0.0006, + "loss": 6.836431980133057, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 7.718832969665527, + "eval_runtime": 58.6615, + "eval_samples_per_second": 41.629, + "eval_steps_per_second": 1.313, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.009791829623281956, + "learning_rate": 0.0006, + "loss": 6.8102593421936035, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.008970947004854679, + "learning_rate": 0.0006, + "loss": 6.792663097381592, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.013689392246305943, + "learning_rate": 0.0006, + "loss": 6.815651893615723, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.031499531120061874, + "learning_rate": 0.0006, + "loss": 6.809841156005859, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.03203094005584717, + "learning_rate": 0.0006, + "loss": 6.779417037963867, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.007947996258735657, + "learning_rate": 0.0006, + "loss": 6.773656845092773, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.03152519091963768, + "learning_rate": 0.0006, + "loss": 6.804083824157715, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.014937258325517178, + "learning_rate": 0.0006, + "loss": 6.792213439941406, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.020961705595254898, + "learning_rate": 0.0006, + "loss": 6.8092803955078125, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.009152988903224468, + "learning_rate": 0.0006, + "loss": 6.7737603187561035, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.02770620957016945, + "learning_rate": 0.0006, + "loss": 6.767014503479004, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.01165871974080801, + "learning_rate": 0.0006, + "loss": 6.760884761810303, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.01910376362502575, + "learning_rate": 0.0006, + "loss": 6.775722980499268, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.010413747280836105, + "learning_rate": 0.0006, + "loss": 6.779816627502441, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.01918417401611805, + "learning_rate": 0.0006, + "loss": 6.760381698608398, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.017960334196686745, + "learning_rate": 0.0006, + "loss": 6.761198997497559, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.016546795144677162, + "learning_rate": 0.0006, + "loss": 6.742448806762695, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.01889597810804844, + "learning_rate": 0.0006, + "loss": 6.758237838745117, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.009611992165446281, + "learning_rate": 0.0006, + "loss": 6.75614070892334, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.02034934051334858, + "learning_rate": 0.0006, + "loss": 6.736965179443359, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.010740065947175026, + "learning_rate": 0.0006, + "loss": 6.746947288513184, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.018383340910077095, + "learning_rate": 0.0006, + "loss": 6.747332572937012, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.012362911365926266, + "learning_rate": 0.0006, + "loss": 6.767669200897217, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.01127463299781084, + "learning_rate": 0.0006, + "loss": 6.769808292388916, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.017693769186735153, + "learning_rate": 0.0006, + "loss": 6.733658790588379, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.006420376244932413, + "learning_rate": 0.0006, + "loss": 6.74934720993042, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.014020364731550217, + "learning_rate": 0.0006, + "loss": 6.728157043457031, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.008018582127988338, + "learning_rate": 0.0006, + "loss": 6.730726718902588, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.00782222580164671, + "learning_rate": 0.0006, + "loss": 6.7499895095825195, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.008278409950435162, + "learning_rate": 0.0006, + "loss": 6.712803363800049, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.009826759807765484, + "learning_rate": 0.0006, + "loss": 6.742262840270996, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.009488224983215332, + "learning_rate": 0.0006, + "loss": 6.739130020141602, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.007667903788387775, + "learning_rate": 0.0006, + "loss": 6.747515678405762, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.0073090000078082085, + "learning_rate": 0.0006, + "loss": 6.740870475769043, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.010058708488941193, + "learning_rate": 0.0006, + "loss": 6.777563095092773, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.020676063373684883, + "learning_rate": 0.0006, + "loss": 6.708603858947754, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.027210544794797897, + "learning_rate": 0.0006, + "loss": 6.694097995758057, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.038053520023822784, + "learning_rate": 0.0006, + "loss": 6.697017669677734, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.04684312269091606, + "learning_rate": 0.0006, + "loss": 6.729883193969727, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.03673060983419418, + "learning_rate": 0.0006, + "loss": 6.743753433227539, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.006882632151246071, + "learning_rate": 0.0006, + "loss": 6.71405029296875, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.03588201105594635, + "learning_rate": 0.0006, + "loss": 6.712467193603516, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.029415199533104897, + "learning_rate": 0.0006, + "loss": 6.715634822845459, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.011425039730966091, + "learning_rate": 0.0006, + "loss": 6.697334289550781, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.0297263003885746, + "learning_rate": 0.0006, + "loss": 6.724848747253418, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.007282314356416464, + "learning_rate": 0.0006, + "loss": 6.702306270599365, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.02673051692545414, + "learning_rate": 0.0006, + "loss": 6.72357177734375, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.016174081712961197, + "learning_rate": 0.0006, + "loss": 6.730698585510254, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.0182416420429945, + "learning_rate": 0.0006, + "loss": 6.717386245727539, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.023250866681337357, + "learning_rate": 0.0006, + "loss": 6.697832107543945, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.007539753336459398, + "learning_rate": 0.0006, + "loss": 6.714348793029785, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.030086630955338478, + "learning_rate": 0.0006, + "loss": 6.724481582641602, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.012784913182258606, + "learning_rate": 0.0006, + "loss": 6.703491687774658, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.02042769268155098, + "learning_rate": 0.0006, + "loss": 6.721139907836914, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.023557329550385475, + "learning_rate": 0.0006, + "loss": 6.728122711181641, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.007716621737927198, + "learning_rate": 0.0006, + "loss": 6.70085334777832, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.022772278636693954, + "learning_rate": 0.0006, + "loss": 6.712882995605469, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.007928933948278427, + "learning_rate": 0.0006, + "loss": 6.698320388793945, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.022355245426297188, + "learning_rate": 0.0006, + "loss": 6.687088489532471, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.02037850394845009, + "learning_rate": 0.0006, + "loss": 6.676856994628906, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.012290362268686295, + "learning_rate": 0.0006, + "loss": 6.682883262634277, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.022331275045871735, + "learning_rate": 0.0006, + "loss": 6.694752216339111, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.007527181413024664, + "learning_rate": 0.0006, + "loss": 6.686394691467285, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.02159409038722515, + "learning_rate": 0.0006, + "loss": 6.697943687438965, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.015054763294756413, + "learning_rate": 0.0006, + "loss": 6.677796363830566, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.01466019731014967, + "learning_rate": 0.0006, + "loss": 6.697715759277344, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.01875465176999569, + "learning_rate": 0.0006, + "loss": 6.700228691101074, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.00778991961851716, + "learning_rate": 0.0006, + "loss": 6.69331693649292, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.020182441920042038, + "learning_rate": 0.0006, + "loss": 6.672651290893555, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.01240516733378172, + "learning_rate": 0.0006, + "loss": 6.688860893249512, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.011218412779271603, + "learning_rate": 0.0006, + "loss": 6.669534683227539, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.0141852255910635, + "learning_rate": 0.0006, + "loss": 6.690206527709961, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 7.537620544433594, + "eval_runtime": 58.4398, + "eval_samples_per_second": 41.787, + "eval_steps_per_second": 1.318, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.008106146939098835, + "learning_rate": 0.0006, + "loss": 6.688591003417969, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.014588244259357452, + "learning_rate": 0.0006, + "loss": 6.684961318969727, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.011985368095338345, + "learning_rate": 0.0006, + "loss": 6.672794342041016, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.014632102102041245, + "learning_rate": 0.0006, + "loss": 6.702736854553223, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.011243713088333607, + "learning_rate": 0.0006, + "loss": 6.658792495727539, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.010519202798604965, + "learning_rate": 0.0006, + "loss": 6.662495136260986, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.007318088319152594, + "learning_rate": 0.0006, + "loss": 6.675954818725586, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.011712398380041122, + "learning_rate": 0.0006, + "loss": 6.673973083496094, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.006726635154336691, + "learning_rate": 0.0006, + "loss": 6.656833171844482, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.005466983187943697, + "learning_rate": 0.0006, + "loss": 6.675468921661377, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.007163621485233307, + "learning_rate": 0.0006, + "loss": 6.6862874031066895, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.007626629434525967, + "learning_rate": 0.0006, + "loss": 6.654741287231445, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.0066330647096037865, + "learning_rate": 0.0006, + "loss": 6.68114709854126, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.007363482844084501, + "learning_rate": 0.0006, + "loss": 6.683980941772461, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.00971233006566763, + "learning_rate": 0.0006, + "loss": 6.669700622558594, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.013185256160795689, + "learning_rate": 0.0006, + "loss": 6.686490535736084, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.019116027280688286, + "learning_rate": 0.0006, + "loss": 6.645479202270508, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.0328814722597599, + "learning_rate": 0.0006, + "loss": 6.68080997467041, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.04482287913560867, + "learning_rate": 0.0006, + "loss": 6.698016166687012, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.03053313121199608, + "learning_rate": 0.0006, + "loss": 6.642240524291992, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.043351784348487854, + "learning_rate": 0.0006, + "loss": 6.662343978881836, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.03165886923670769, + "learning_rate": 0.0006, + "loss": 6.6291961669921875, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.028601787984371185, + "learning_rate": 0.0006, + "loss": 6.667544364929199, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.019558480009436607, + "learning_rate": 0.0006, + "loss": 6.641025543212891, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.039851363748311996, + "learning_rate": 0.0006, + "loss": 6.665952205657959, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.018417146056890488, + "learning_rate": 0.0006, + "loss": 6.657257556915283, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.02894527092576027, + "learning_rate": 0.0006, + "loss": 6.656288146972656, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.01910966821014881, + "learning_rate": 0.0006, + "loss": 6.688958168029785, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.01668412610888481, + "learning_rate": 0.0006, + "loss": 6.64876127243042, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.02935275249183178, + "learning_rate": 0.0006, + "loss": 6.626096725463867, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.017997972667217255, + "learning_rate": 0.0006, + "loss": 6.640386581420898, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.01779189519584179, + "learning_rate": 0.0006, + "loss": 6.641165256500244, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.013901712372899055, + "learning_rate": 0.0006, + "loss": 6.651627063751221, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013211199082434177, + "learning_rate": 0.0006, + "loss": 6.630064964294434, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.02324107475578785, + "learning_rate": 0.0006, + "loss": 6.670443058013916, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.019452504813671112, + "learning_rate": 0.0006, + "loss": 6.651266098022461, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.017943572252988815, + "learning_rate": 0.0006, + "loss": 6.657943248748779, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.012136237695813179, + "learning_rate": 0.0006, + "loss": 6.640082359313965, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.011544011533260345, + "learning_rate": 0.0006, + "loss": 6.6186933517456055, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.01569426618516445, + "learning_rate": 0.0006, + "loss": 6.640767574310303, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.027668394148349762, + "learning_rate": 0.0006, + "loss": 6.651412487030029, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.04073230177164078, + "learning_rate": 0.0006, + "loss": 6.636565208435059, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.04942672327160835, + "learning_rate": 0.0006, + "loss": 6.667994499206543, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.037568598985672, + "learning_rate": 0.0006, + "loss": 6.632008075714111, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.009880400262773037, + "learning_rate": 0.0006, + "loss": 6.6504058837890625, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.033384259790182114, + "learning_rate": 0.0006, + "loss": 6.63852596282959, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.035640884190797806, + "learning_rate": 0.0006, + "loss": 6.627986431121826, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.013637864030897617, + "learning_rate": 0.0006, + "loss": 6.624503135681152, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.025304675102233887, + "learning_rate": 0.0006, + "loss": 6.612077236175537, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.03377160802483559, + "learning_rate": 0.0006, + "loss": 6.608938217163086, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.019803153350949287, + "learning_rate": 0.0006, + "loss": 6.636343955993652, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.018928129225969315, + "learning_rate": 0.0006, + "loss": 6.636613845825195, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.023068996146321297, + "learning_rate": 0.0006, + "loss": 6.6258440017700195, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.016849957406520844, + "learning_rate": 0.0006, + "loss": 6.638483047485352, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.021430544555187225, + "learning_rate": 0.0006, + "loss": 6.619649410247803, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.024053368717432022, + "learning_rate": 0.0006, + "loss": 6.639214515686035, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.011926693841814995, + "learning_rate": 0.0006, + "loss": 6.604836463928223, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.01961611397564411, + "learning_rate": 0.0006, + "loss": 6.624864101409912, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.027946067973971367, + "learning_rate": 0.0006, + "loss": 6.621284484863281, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.02108592912554741, + "learning_rate": 0.0006, + "loss": 6.5982465744018555, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.017272260040044785, + "learning_rate": 0.0006, + "loss": 6.626142501831055, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.015654582530260086, + "learning_rate": 0.0006, + "loss": 6.621735572814941, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.015797540545463562, + "learning_rate": 0.0006, + "loss": 6.6016411781311035, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.028488904237747192, + "learning_rate": 0.0006, + "loss": 6.598413467407227, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.03411734476685524, + "learning_rate": 0.0006, + "loss": 6.597321033477783, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.027188263833522797, + "learning_rate": 0.0006, + "loss": 6.630616188049316, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.03411437198519707, + "learning_rate": 0.0006, + "loss": 6.57647705078125, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.019907327368855476, + "learning_rate": 0.0006, + "loss": 6.579545974731445, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.02485930174589157, + "learning_rate": 0.0006, + "loss": 6.591073036193848, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.026106052100658417, + "learning_rate": 0.0006, + "loss": 6.580729961395264, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.029401391744613647, + "learning_rate": 0.0006, + "loss": 6.5460076332092285, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.032869379967451096, + "learning_rate": 0.0006, + "loss": 6.583278179168701, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 7.416288375854492, + "eval_runtime": 60.4177, + "eval_samples_per_second": 40.419, + "eval_steps_per_second": 1.274, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.04862433671951294, + "learning_rate": 0.0006, + "loss": 6.58473014831543, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.06636223942041397, + "learning_rate": 0.0006, + "loss": 6.621153354644775, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.05888264626264572, + "learning_rate": 0.0006, + "loss": 6.621196269989014, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.05331697314977646, + "learning_rate": 0.0006, + "loss": 6.600763320922852, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.03947671875357628, + "learning_rate": 0.0006, + "loss": 6.600671768188477, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.04045343026518822, + "learning_rate": 0.0006, + "loss": 6.63529109954834, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.029766036197543144, + "learning_rate": 0.0006, + "loss": 6.612868309020996, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.03416327014565468, + "learning_rate": 0.0006, + "loss": 6.582240581512451, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.03458515554666519, + "learning_rate": 0.0006, + "loss": 6.604381561279297, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.024099089205265045, + "learning_rate": 0.0006, + "loss": 6.597355365753174, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.02694438211619854, + "learning_rate": 0.0006, + "loss": 6.597512245178223, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.024201856926083565, + "learning_rate": 0.0006, + "loss": 6.564685821533203, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.015959380194544792, + "learning_rate": 0.0006, + "loss": 6.582263946533203, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.025690171867609024, + "learning_rate": 0.0006, + "loss": 6.601057052612305, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.014497052878141403, + "learning_rate": 0.0006, + "loss": 6.584796905517578, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.019870297983288765, + "learning_rate": 0.0006, + "loss": 6.574010372161865, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.01807512901723385, + "learning_rate": 0.0006, + "loss": 6.558596134185791, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.014525193721055984, + "learning_rate": 0.0006, + "loss": 6.571803092956543, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.016113435849547386, + "learning_rate": 0.0006, + "loss": 6.568970680236816, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.020724017173051834, + "learning_rate": 0.0006, + "loss": 6.5669403076171875, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.013963675126433372, + "learning_rate": 0.0006, + "loss": 6.586382865905762, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.01621115580201149, + "learning_rate": 0.0006, + "loss": 6.539546966552734, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.014925207942724228, + "learning_rate": 0.0006, + "loss": 6.531795501708984, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.015454350039362907, + "learning_rate": 0.0006, + "loss": 6.575699806213379, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.01623542234301567, + "learning_rate": 0.0006, + "loss": 6.582373142242432, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015452183783054352, + "learning_rate": 0.0006, + "loss": 6.576451301574707, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.021102432161569595, + "learning_rate": 0.0006, + "loss": 6.543055534362793, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.009827060624957085, + "learning_rate": 0.0006, + "loss": 6.526939392089844, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.013735136948525906, + "learning_rate": 0.0006, + "loss": 6.538825988769531, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02073509432375431, + "learning_rate": 0.0006, + "loss": 6.521315574645996, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.03110959194600582, + "learning_rate": 0.0006, + "loss": 6.561807632446289, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.045058365911245346, + "learning_rate": 0.0006, + "loss": 6.526515960693359, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.06359834969043732, + "learning_rate": 0.0006, + "loss": 6.554903030395508, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.06406767666339874, + "learning_rate": 0.0006, + "loss": 6.576272010803223, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.03403699770569801, + "learning_rate": 0.0006, + "loss": 6.5306243896484375, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.036109618842601776, + "learning_rate": 0.0006, + "loss": 6.54176139831543, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.03478201478719711, + "learning_rate": 0.0006, + "loss": 6.530734062194824, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.03401544317603111, + "learning_rate": 0.0006, + "loss": 6.557794094085693, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.031629662960767746, + "learning_rate": 0.0006, + "loss": 6.564887046813965, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.02305024117231369, + "learning_rate": 0.0006, + "loss": 6.547412872314453, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.030852800235152245, + "learning_rate": 0.0006, + "loss": 6.495370388031006, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.03368550166487694, + "learning_rate": 0.0006, + "loss": 6.495980262756348, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.020976359024643898, + "learning_rate": 0.0006, + "loss": 6.509410858154297, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.019652143120765686, + "learning_rate": 0.0006, + "loss": 6.493232727050781, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.024320121854543686, + "learning_rate": 0.0006, + "loss": 6.521531581878662, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.016611427068710327, + "learning_rate": 0.0006, + "loss": 6.507271766662598, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01635856367647648, + "learning_rate": 0.0006, + "loss": 6.492038726806641, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.017642095685005188, + "learning_rate": 0.0006, + "loss": 6.500349998474121, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.01679178513586521, + "learning_rate": 0.0006, + "loss": 6.500361442565918, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.011776307597756386, + "learning_rate": 0.0006, + "loss": 6.482490539550781, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.0176856629550457, + "learning_rate": 0.0006, + "loss": 6.48488712310791, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.015193155966699123, + "learning_rate": 0.0006, + "loss": 6.459545612335205, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.018623212352395058, + "learning_rate": 0.0006, + "loss": 6.478870391845703, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.012841006740927696, + "learning_rate": 0.0006, + "loss": 6.474418640136719, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.015540511347353458, + "learning_rate": 0.0006, + "loss": 6.4706525802612305, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.022695740684866905, + "learning_rate": 0.0006, + "loss": 6.455738067626953, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.026930129155516624, + "learning_rate": 0.0006, + "loss": 6.458016872406006, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.03555797412991524, + "learning_rate": 0.0006, + "loss": 6.480044364929199, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.05024682730436325, + "learning_rate": 0.0006, + "loss": 6.454564571380615, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.06616255640983582, + "learning_rate": 0.0006, + "loss": 6.433927536010742, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.059065770357847214, + "learning_rate": 0.0006, + "loss": 6.508362770080566, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.026689983904361725, + "learning_rate": 0.0006, + "loss": 6.464297294616699, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.04616398364305496, + "learning_rate": 0.0006, + "loss": 6.45448112487793, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.04980356618762016, + "learning_rate": 0.0006, + "loss": 6.4696574211120605, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.0551983006298542, + "learning_rate": 0.0006, + "loss": 6.419000625610352, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.06988243758678436, + "learning_rate": 0.0006, + "loss": 6.442890167236328, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.06040226295590401, + "learning_rate": 0.0006, + "loss": 6.513258934020996, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.04635024070739746, + "learning_rate": 0.0006, + "loss": 6.4710283279418945, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.04762702435255051, + "learning_rate": 0.0006, + "loss": 6.469327926635742, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.04857536032795906, + "learning_rate": 0.0006, + "loss": 6.426105499267578, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.04412693530321121, + "learning_rate": 0.0006, + "loss": 6.43179988861084, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.026129672303795815, + "learning_rate": 0.0006, + "loss": 6.437918663024902, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 7.219249248504639, + "eval_runtime": 58.6837, + "eval_samples_per_second": 41.613, + "eval_steps_per_second": 1.312, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.03329874947667122, + "learning_rate": 0.0006, + "loss": 6.438179969787598, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.03723187372088432, + "learning_rate": 0.0006, + "loss": 6.394242286682129, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.014625726267695427, + "learning_rate": 0.0006, + "loss": 6.412118911743164, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.02917475439608097, + "learning_rate": 0.0006, + "loss": 6.453164577484131, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01811208389699459, + "learning_rate": 0.0006, + "loss": 6.396003246307373, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.027235351502895355, + "learning_rate": 0.0006, + "loss": 6.402976036071777, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.02835794910788536, + "learning_rate": 0.0006, + "loss": 6.380568504333496, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.020868154242634773, + "learning_rate": 0.0006, + "loss": 6.38240909576416, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.02113114856183529, + "learning_rate": 0.0006, + "loss": 6.377391815185547, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.018894119188189507, + "learning_rate": 0.0006, + "loss": 6.354569435119629, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.01820225641131401, + "learning_rate": 0.0006, + "loss": 6.392882347106934, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.017510728910565376, + "learning_rate": 0.0006, + "loss": 6.364162445068359, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.017893686890602112, + "learning_rate": 0.0006, + "loss": 6.373456954956055, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.013382102362811565, + "learning_rate": 0.0006, + "loss": 6.3714399337768555, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.022540833801031113, + "learning_rate": 0.0006, + "loss": 6.348170757293701, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.02110104262828827, + "learning_rate": 0.0006, + "loss": 6.413206100463867, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.014498349279165268, + "learning_rate": 0.0006, + "loss": 6.361384391784668, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.01819637045264244, + "learning_rate": 0.0006, + "loss": 6.315275192260742, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.0195972491055727, + "learning_rate": 0.0006, + "loss": 6.321994781494141, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.018822453916072845, + "learning_rate": 0.0006, + "loss": 6.345219612121582, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.019247712567448616, + "learning_rate": 0.0006, + "loss": 6.348073959350586, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.03175375238060951, + "learning_rate": 0.0006, + "loss": 6.300344944000244, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.06326444447040558, + "learning_rate": 0.0006, + "loss": 6.392579078674316, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.09667232632637024, + "learning_rate": 0.0006, + "loss": 6.36024284362793, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.05774446204304695, + "learning_rate": 0.0006, + "loss": 6.365338325500488, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.08793356269598007, + "learning_rate": 0.0006, + "loss": 6.43284273147583, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.0640595331788063, + "learning_rate": 0.0006, + "loss": 6.342428207397461, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.0447838194668293, + "learning_rate": 0.0006, + "loss": 6.367971897125244, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.04768967255949974, + "learning_rate": 0.0006, + "loss": 6.35027551651001, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.027981851249933243, + "learning_rate": 0.0006, + "loss": 6.33723258972168, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.03479623422026634, + "learning_rate": 0.0006, + "loss": 6.36590051651001, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.020846687257289886, + "learning_rate": 0.0006, + "loss": 6.315143585205078, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.0312928780913353, + "learning_rate": 0.0006, + "loss": 6.340150833129883, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.020593881607055664, + "learning_rate": 0.0006, + "loss": 6.348175048828125, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.022010236978530884, + "learning_rate": 0.0006, + "loss": 6.318607330322266, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.018427059054374695, + "learning_rate": 0.0006, + "loss": 6.331822395324707, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.019964968785643578, + "learning_rate": 0.0006, + "loss": 6.287320137023926, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.02130270190536976, + "learning_rate": 0.0006, + "loss": 6.353344917297363, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.01616514101624489, + "learning_rate": 0.0006, + "loss": 6.314052581787109, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.016639474779367447, + "learning_rate": 0.0006, + "loss": 6.320351600646973, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.017620692029595375, + "learning_rate": 0.0006, + "loss": 6.311459064483643, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.013542844913899899, + "learning_rate": 0.0006, + "loss": 6.324368476867676, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.01308678649365902, + "learning_rate": 0.0006, + "loss": 6.307419776916504, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.014672392047941685, + "learning_rate": 0.0006, + "loss": 6.2785420417785645, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01399585697799921, + "learning_rate": 0.0006, + "loss": 6.258619785308838, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.01330904383212328, + "learning_rate": 0.0006, + "loss": 6.299595832824707, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.016613377258181572, + "learning_rate": 0.0006, + "loss": 6.2900309562683105, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.015076331794261932, + "learning_rate": 0.0006, + "loss": 6.231028079986572, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.01589728519320488, + "learning_rate": 0.0006, + "loss": 6.257922172546387, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.030972328037023544, + "learning_rate": 0.0006, + "loss": 6.285238265991211, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.05048783868551254, + "learning_rate": 0.0006, + "loss": 6.287471771240234, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.07390714436769485, + "learning_rate": 0.0006, + "loss": 6.284293174743652, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.07166943699121475, + "learning_rate": 0.0006, + "loss": 6.290704727172852, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.028792452067136765, + "learning_rate": 0.0006, + "loss": 6.243415832519531, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.04822041466832161, + "learning_rate": 0.0006, + "loss": 6.278229713439941, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.036688923835754395, + "learning_rate": 0.0006, + "loss": 6.275590896606445, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.03679114952683449, + "learning_rate": 0.0006, + "loss": 6.236509323120117, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.023249102756381035, + "learning_rate": 0.0006, + "loss": 6.230788230895996, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.02682286500930786, + "learning_rate": 0.0006, + "loss": 6.278890609741211, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.02202044241130352, + "learning_rate": 0.0006, + "loss": 6.234531402587891, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.02545459009706974, + "learning_rate": 0.0006, + "loss": 6.271790981292725, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.02012932486832142, + "learning_rate": 0.0006, + "loss": 6.247462749481201, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.02216474711894989, + "learning_rate": 0.0006, + "loss": 6.26396369934082, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.016078704968094826, + "learning_rate": 0.0006, + "loss": 6.228609085083008, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.023021478205919266, + "learning_rate": 0.0006, + "loss": 6.197244167327881, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.024766333401203156, + "learning_rate": 0.0006, + "loss": 6.181090354919434, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.025581466034054756, + "learning_rate": 0.0006, + "loss": 6.201817512512207, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.016964981332421303, + "learning_rate": 0.0006, + "loss": 6.201486587524414, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.023957721889019012, + "learning_rate": 0.0006, + "loss": 6.185701847076416, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.029775245115160942, + "learning_rate": 0.0006, + "loss": 6.242060661315918, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.04007191210985184, + "learning_rate": 0.0006, + "loss": 6.19931697845459, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.04882276430726051, + "learning_rate": 0.0006, + "loss": 6.1880598068237305, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 7.018043518066406, + "eval_runtime": 59.8815, + "eval_samples_per_second": 40.781, + "eval_steps_per_second": 1.286, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.07144106924533844, + "learning_rate": 0.0006, + "loss": 6.172799110412598, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.06712841987609863, + "learning_rate": 0.0006, + "loss": 6.228682518005371, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.04405239224433899, + "learning_rate": 0.0006, + "loss": 6.1827898025512695, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.03418995440006256, + "learning_rate": 0.0006, + "loss": 6.200876235961914, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.04002382606267929, + "learning_rate": 0.0006, + "loss": 6.1799211502075195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.029748762026429176, + "learning_rate": 0.0006, + "loss": 6.125422477722168, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.03408357501029968, + "learning_rate": 0.0006, + "loss": 6.203458309173584, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.02227279357612133, + "learning_rate": 0.0006, + "loss": 6.135439872741699, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02727862447500229, + "learning_rate": 0.0006, + "loss": 6.202380180358887, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.026019075885415077, + "learning_rate": 0.0006, + "loss": 6.201343536376953, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.027227282524108887, + "learning_rate": 0.0006, + "loss": 6.164059638977051, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.027649980038404465, + "learning_rate": 0.0006, + "loss": 6.135448932647705, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.02168039232492447, + "learning_rate": 0.0006, + "loss": 6.154831886291504, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.02101467363536358, + "learning_rate": 0.0006, + "loss": 6.128695487976074, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.018344245851039886, + "learning_rate": 0.0006, + "loss": 6.136524200439453, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.020323453471064568, + "learning_rate": 0.0006, + "loss": 6.1206488609313965, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.01750849559903145, + "learning_rate": 0.0006, + "loss": 6.128170967102051, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.017778221517801285, + "learning_rate": 0.0006, + "loss": 6.150458812713623, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.01863938942551613, + "learning_rate": 0.0006, + "loss": 6.100106716156006, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.020196080207824707, + "learning_rate": 0.0006, + "loss": 6.149346351623535, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.03203890845179558, + "learning_rate": 0.0006, + "loss": 6.124648094177246, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.05325170233845711, + "learning_rate": 0.0006, + "loss": 6.115618705749512, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.05888345465064049, + "learning_rate": 0.0006, + "loss": 6.165221214294434, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.026873115450143814, + "learning_rate": 0.0006, + "loss": 6.039590835571289, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.06304649263620377, + "learning_rate": 0.0006, + "loss": 6.131900787353516, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.08567557483911514, + "learning_rate": 0.0006, + "loss": 6.218114852905273, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.08868218213319778, + "learning_rate": 0.0006, + "loss": 6.216706275939941, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.03955601155757904, + "learning_rate": 0.0006, + "loss": 6.1566925048828125, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.04979498311877251, + "learning_rate": 0.0006, + "loss": 6.19327449798584, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.03253026679158211, + "learning_rate": 0.0006, + "loss": 6.133993148803711, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.03016652725636959, + "learning_rate": 0.0006, + "loss": 6.174962997436523, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.03159862384200096, + "learning_rate": 0.0006, + "loss": 6.082124710083008, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.027673959732055664, + "learning_rate": 0.0006, + "loss": 6.111873149871826, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.022789714857935905, + "learning_rate": 0.0006, + "loss": 6.129437446594238, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.030061012133955956, + "learning_rate": 0.0006, + "loss": 6.126370429992676, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.019007068127393723, + "learning_rate": 0.0006, + "loss": 6.07887077331543, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.01897653006017208, + "learning_rate": 0.0006, + "loss": 6.096936225891113, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.024659698829054832, + "learning_rate": 0.0006, + "loss": 6.0933380126953125, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.021924549713730812, + "learning_rate": 0.0006, + "loss": 6.105311870574951, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.018817558884620667, + "learning_rate": 0.0006, + "loss": 6.058895111083984, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.02477840520441532, + "learning_rate": 0.0006, + "loss": 6.038928031921387, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.018968632444739342, + "learning_rate": 0.0006, + "loss": 6.1368937492370605, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.02409801445901394, + "learning_rate": 0.0006, + "loss": 6.110324859619141, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.017595326527953148, + "learning_rate": 0.0006, + "loss": 6.074368476867676, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.01875406503677368, + "learning_rate": 0.0006, + "loss": 6.121023654937744, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.018781628459692, + "learning_rate": 0.0006, + "loss": 6.064914703369141, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.017003126442432404, + "learning_rate": 0.0006, + "loss": 6.039608001708984, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.017665037885308266, + "learning_rate": 0.0006, + "loss": 6.053671836853027, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.02066265605390072, + "learning_rate": 0.0006, + "loss": 6.090536117553711, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.02888769656419754, + "learning_rate": 0.0006, + "loss": 5.9858903884887695, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.03744485229253769, + "learning_rate": 0.0006, + "loss": 6.086264610290527, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.04719431325793266, + "learning_rate": 0.0006, + "loss": 6.09747314453125, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.044963907450437546, + "learning_rate": 0.0006, + "loss": 6.091272354125977, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.03810194507241249, + "learning_rate": 0.0006, + "loss": 6.084904193878174, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.04626769199967384, + "learning_rate": 0.0006, + "loss": 6.017384052276611, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.03459552302956581, + "learning_rate": 0.0006, + "loss": 6.111314296722412, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.027385666966438293, + "learning_rate": 0.0006, + "loss": 6.002509117126465, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.03417652100324631, + "learning_rate": 0.0006, + "loss": 6.0421061515808105, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.04415450245141983, + "learning_rate": 0.0006, + "loss": 6.035305023193359, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.03259635716676712, + "learning_rate": 0.0006, + "loss": 6.025840759277344, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.024941129609942436, + "learning_rate": 0.0006, + "loss": 6.043416976928711, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.02434500865638256, + "learning_rate": 0.0006, + "loss": 6.002052307128906, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.01876584254205227, + "learning_rate": 0.0006, + "loss": 6.045035362243652, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.02254367619752884, + "learning_rate": 0.0006, + "loss": 5.972503662109375, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.024504603818058968, + "learning_rate": 0.0006, + "loss": 5.992846488952637, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.02317827194929123, + "learning_rate": 0.0006, + "loss": 6.007697582244873, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02225821278989315, + "learning_rate": 0.0006, + "loss": 6.01463508605957, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.01811717264354229, + "learning_rate": 0.0006, + "loss": 5.995022773742676, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.019589431583881378, + "learning_rate": 0.0006, + "loss": 5.990633964538574, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.015422016382217407, + "learning_rate": 0.0006, + "loss": 6.011263847351074, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.023298880085349083, + "learning_rate": 0.0006, + "loss": 5.9954938888549805, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.022172044962644577, + "learning_rate": 0.0006, + "loss": 5.95679235458374, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 6.75809383392334, + "eval_runtime": 72.8926, + "eval_samples_per_second": 33.501, + "eval_steps_per_second": 1.056, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.029082221910357475, + "learning_rate": 0.0006, + "loss": 6.018675804138184, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.03959495946764946, + "learning_rate": 0.0006, + "loss": 5.93310546875, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.043756261467933655, + "learning_rate": 0.0006, + "loss": 5.965944290161133, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.04996631667017937, + "learning_rate": 0.0006, + "loss": 6.0381059646606445, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.05721999332308769, + "learning_rate": 0.0006, + "loss": 5.968381881713867, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.07594525814056396, + "learning_rate": 0.0006, + "loss": 6.027225494384766, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.06861685961484909, + "learning_rate": 0.0006, + "loss": 6.035175323486328, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.04156472906470299, + "learning_rate": 0.0006, + "loss": 5.9500017166137695, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.045399412512779236, + "learning_rate": 0.0006, + "loss": 5.9953389167785645, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.04399585723876953, + "learning_rate": 0.0006, + "loss": 5.986051559448242, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.05764509737491608, + "learning_rate": 0.0006, + "loss": 6.005392074584961, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.060871850699186325, + "learning_rate": 0.0006, + "loss": 6.01304817199707, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.03381241112947464, + "learning_rate": 0.0006, + "loss": 5.977019309997559, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.03384104371070862, + "learning_rate": 0.0006, + "loss": 5.988002777099609, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.03120286948978901, + "learning_rate": 0.0006, + "loss": 5.968227386474609, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.026532653719186783, + "learning_rate": 0.0006, + "loss": 5.936810493469238, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.024956922978162766, + "learning_rate": 0.0006, + "loss": 5.9348015785217285, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.025926144793629646, + "learning_rate": 0.0006, + "loss": 5.952630519866943, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.02893793396651745, + "learning_rate": 0.0006, + "loss": 5.9569854736328125, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.0320001058280468, + "learning_rate": 0.0006, + "loss": 5.917050838470459, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.03925463184714317, + "learning_rate": 0.0006, + "loss": 5.9925127029418945, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.05096640810370445, + "learning_rate": 0.0006, + "loss": 5.926358222961426, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.041637953370809555, + "learning_rate": 0.0006, + "loss": 5.954983711242676, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.03138125315308571, + "learning_rate": 0.0006, + "loss": 5.893692970275879, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.029299303889274597, + "learning_rate": 0.0006, + "loss": 5.94670295715332, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.023969192057847977, + "learning_rate": 0.0006, + "loss": 5.925658702850342, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.025406932458281517, + "learning_rate": 0.0006, + "loss": 5.966875076293945, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.024309059605002403, + "learning_rate": 0.0006, + "loss": 5.932581424713135, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.017237074673175812, + "learning_rate": 0.0006, + "loss": 5.898823261260986, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.025594931095838547, + "learning_rate": 0.0006, + "loss": 5.9285993576049805, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.025184160098433495, + "learning_rate": 0.0006, + "loss": 5.915966510772705, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.019004398956894875, + "learning_rate": 0.0006, + "loss": 5.893395900726318, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.021802261471748352, + "learning_rate": 0.0006, + "loss": 5.953311443328857, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.02202655002474785, + "learning_rate": 0.0006, + "loss": 5.953000545501709, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.02067241445183754, + "learning_rate": 0.0006, + "loss": 5.869247913360596, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.015430927276611328, + "learning_rate": 0.0006, + "loss": 5.855883598327637, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.019917506724596024, + "learning_rate": 0.0006, + "loss": 5.9284539222717285, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.024291323497891426, + "learning_rate": 0.0006, + "loss": 5.973604202270508, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.022133296355605125, + "learning_rate": 0.0006, + "loss": 5.85693359375, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0254366435110569, + "learning_rate": 0.0006, + "loss": 5.939558982849121, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.030134694650769234, + "learning_rate": 0.0006, + "loss": 5.892992973327637, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.03845192864537239, + "learning_rate": 0.0006, + "loss": 5.89303731918335, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.042409464716911316, + "learning_rate": 0.0006, + "loss": 5.914068222045898, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.0454271174967289, + "learning_rate": 0.0006, + "loss": 5.853579044342041, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.04934931918978691, + "learning_rate": 0.0006, + "loss": 5.881861686706543, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.0519695058465004, + "learning_rate": 0.0006, + "loss": 5.872620582580566, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.06172563508152962, + "learning_rate": 0.0006, + "loss": 5.938283443450928, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.08416368812322617, + "learning_rate": 0.0006, + "loss": 5.943403720855713, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.07311230152845383, + "learning_rate": 0.0006, + "loss": 5.954029083251953, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.09851036965847015, + "learning_rate": 0.0006, + "loss": 5.992746829986572, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.07521767914295197, + "learning_rate": 0.0006, + "loss": 5.956884384155273, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.05720978602766991, + "learning_rate": 0.0006, + "loss": 5.911205291748047, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.05360936373472214, + "learning_rate": 0.0006, + "loss": 5.9734787940979, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.035826463252305984, + "learning_rate": 0.0006, + "loss": 5.95751953125, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.030137093737721443, + "learning_rate": 0.0006, + "loss": 5.91911506652832, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.030039150267839432, + "learning_rate": 0.0006, + "loss": 5.910290718078613, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.03349437564611435, + "learning_rate": 0.0006, + "loss": 5.904623985290527, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.02211189828813076, + "learning_rate": 0.0006, + "loss": 5.909668445587158, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.022601254284381866, + "learning_rate": 0.0006, + "loss": 5.942885875701904, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.02197270467877388, + "learning_rate": 0.0006, + "loss": 5.867103576660156, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.022765018045902252, + "learning_rate": 0.0006, + "loss": 5.894528865814209, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.021181151270866394, + "learning_rate": 0.0006, + "loss": 5.896910667419434, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.01878993771970272, + "learning_rate": 0.0006, + "loss": 5.860502243041992, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.014679803512990475, + "learning_rate": 0.0006, + "loss": 5.8725266456604, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.017919695004820824, + "learning_rate": 0.0006, + "loss": 5.88902473449707, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.01437009871006012, + "learning_rate": 0.0006, + "loss": 5.868186950683594, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.014293290674686432, + "learning_rate": 0.0006, + "loss": 5.833364963531494, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.015440084971487522, + "learning_rate": 0.0006, + "loss": 5.860260009765625, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.014292514882981777, + "learning_rate": 0.0006, + "loss": 5.844939231872559, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.015702251344919205, + "learning_rate": 0.0006, + "loss": 5.825929164886475, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.013372798450291157, + "learning_rate": 0.0006, + "loss": 5.83062744140625, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.013794616796076298, + "learning_rate": 0.0006, + "loss": 5.7897820472717285, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 6.600830078125, + "eval_runtime": 71.4316, + "eval_samples_per_second": 34.187, + "eval_steps_per_second": 1.078, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.01262663770467043, + "learning_rate": 0.0006, + "loss": 5.802824020385742, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.016375157982110977, + "learning_rate": 0.0006, + "loss": 5.8018317222595215, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.010541851632297039, + "learning_rate": 0.0006, + "loss": 5.848208427429199, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.011726257391273975, + "learning_rate": 0.0006, + "loss": 5.771823883056641, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01124380249530077, + "learning_rate": 0.0006, + "loss": 5.773855209350586, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.010801885277032852, + "learning_rate": 0.0006, + "loss": 5.847390651702881, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.013018375262618065, + "learning_rate": 0.0006, + "loss": 5.786786079406738, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.014233003370463848, + "learning_rate": 0.0006, + "loss": 5.7913713455200195, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.01465831883251667, + "learning_rate": 0.0006, + "loss": 5.821170330047607, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.017383430153131485, + "learning_rate": 0.0006, + "loss": 5.811469078063965, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.020328007638454437, + "learning_rate": 0.0006, + "loss": 5.718976020812988, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.01776740700006485, + "learning_rate": 0.0006, + "loss": 5.80593729019165, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.017713800072669983, + "learning_rate": 0.0006, + "loss": 5.744930267333984, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.028774067759513855, + "learning_rate": 0.0006, + "loss": 5.85262393951416, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.05153049901127815, + "learning_rate": 0.0006, + "loss": 5.739948272705078, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.06882233172655106, + "learning_rate": 0.0006, + "loss": 5.844270706176758, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.0420759879052639, + "learning_rate": 0.0006, + "loss": 5.79060173034668, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.05598323419690132, + "learning_rate": 0.0006, + "loss": 5.833735466003418, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.05011717975139618, + "learning_rate": 0.0006, + "loss": 5.840887069702148, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.040782492607831955, + "learning_rate": 0.0006, + "loss": 5.7905449867248535, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.030804185196757317, + "learning_rate": 0.0006, + "loss": 5.802691459655762, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.032001759856939316, + "learning_rate": 0.0006, + "loss": 5.726612091064453, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.0295492485165596, + "learning_rate": 0.0006, + "loss": 5.798826694488525, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.034406546503305435, + "learning_rate": 0.0006, + "loss": 5.720219135284424, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.02627669647336006, + "learning_rate": 0.0006, + "loss": 5.723702430725098, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.0282363910228014, + "learning_rate": 0.0006, + "loss": 5.818911552429199, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.017894433811306953, + "learning_rate": 0.0006, + "loss": 5.772168159484863, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.019017303362488747, + "learning_rate": 0.0006, + "loss": 5.774587631225586, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.022159283980727196, + "learning_rate": 0.0006, + "loss": 5.812951564788818, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02198116108775139, + "learning_rate": 0.0006, + "loss": 5.808734893798828, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.017741188406944275, + "learning_rate": 0.0006, + "loss": 5.776142120361328, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.017197400331497192, + "learning_rate": 0.0006, + "loss": 5.7471208572387695, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.01752566732466221, + "learning_rate": 0.0006, + "loss": 5.780961990356445, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.017961280420422554, + "learning_rate": 0.0006, + "loss": 5.713686943054199, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.02227521874010563, + "learning_rate": 0.0006, + "loss": 5.717774868011475, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.024075504392385483, + "learning_rate": 0.0006, + "loss": 5.715015411376953, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.029195033013820648, + "learning_rate": 0.0006, + "loss": 5.659659385681152, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.02959120273590088, + "learning_rate": 0.0006, + "loss": 5.78981351852417, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.021854368969798088, + "learning_rate": 0.0006, + "loss": 5.750844955444336, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.017120081931352615, + "learning_rate": 0.0006, + "loss": 5.672736167907715, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.020712822675704956, + "learning_rate": 0.0006, + "loss": 5.708046913146973, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.019186170771718025, + "learning_rate": 0.0006, + "loss": 5.71742057800293, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.01985814794898033, + "learning_rate": 0.0006, + "loss": 5.735311508178711, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.01965234987437725, + "learning_rate": 0.0006, + "loss": 5.728727340698242, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.019827688112854958, + "learning_rate": 0.0006, + "loss": 5.743068695068359, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.025500580668449402, + "learning_rate": 0.0006, + "loss": 5.76585578918457, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.03701791539788246, + "learning_rate": 0.0006, + "loss": 5.728367805480957, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.04173659533262253, + "learning_rate": 0.0006, + "loss": 5.683089256286621, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.031390558928251266, + "learning_rate": 0.0006, + "loss": 5.740921974182129, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.03970158472657204, + "learning_rate": 0.0006, + "loss": 5.730632781982422, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.03865141049027443, + "learning_rate": 0.0006, + "loss": 5.686657905578613, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.04856107011437416, + "learning_rate": 0.0006, + "loss": 5.657513618469238, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.04016133397817612, + "learning_rate": 0.0006, + "loss": 5.754552841186523, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.03961458429694176, + "learning_rate": 0.0006, + "loss": 5.672088623046875, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.03403536602854729, + "learning_rate": 0.0006, + "loss": 5.736673355102539, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.03766333684325218, + "learning_rate": 0.0006, + "loss": 5.716219902038574, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.04085501655936241, + "learning_rate": 0.0006, + "loss": 5.678958415985107, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.036888182163238525, + "learning_rate": 0.0006, + "loss": 5.6990461349487305, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.0492791011929512, + "learning_rate": 0.0006, + "loss": 5.78053092956543, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.05572429671883583, + "learning_rate": 0.0006, + "loss": 5.7543535232543945, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.047000352293252945, + "learning_rate": 0.0006, + "loss": 5.7174272537231445, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.058798179030418396, + "learning_rate": 0.0006, + "loss": 5.793713569641113, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.046992287039756775, + "learning_rate": 0.0006, + "loss": 5.784457683563232, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.04382370039820671, + "learning_rate": 0.0006, + "loss": 5.793731212615967, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.040616173297166824, + "learning_rate": 0.0006, + "loss": 5.69599723815918, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.03446170687675476, + "learning_rate": 0.0006, + "loss": 5.694674968719482, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.029936550185084343, + "learning_rate": 0.0006, + "loss": 5.6704487800598145, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02224753610789776, + "learning_rate": 0.0006, + "loss": 5.728894233703613, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.024346651509404182, + "learning_rate": 0.0006, + "loss": 5.772412300109863, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.022997068241238594, + "learning_rate": 0.0006, + "loss": 5.645310401916504, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.024918686598539352, + "learning_rate": 0.0006, + "loss": 5.671009063720703, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.020067067816853523, + "learning_rate": 0.0006, + "loss": 5.637441635131836, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 6.387807846069336, + "eval_runtime": 72.3372, + "eval_samples_per_second": 33.759, + "eval_steps_per_second": 1.064, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02015046589076519, + "learning_rate": 0.0006, + "loss": 5.598219871520996, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.017398346215486526, + "learning_rate": 0.0006, + "loss": 5.66834831237793, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.016871167346835136, + "learning_rate": 0.0006, + "loss": 5.673916816711426, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.01695060543715954, + "learning_rate": 0.0006, + "loss": 5.646677017211914, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.019298404455184937, + "learning_rate": 0.0006, + "loss": 5.663064479827881, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018699361011385918, + "learning_rate": 0.0006, + "loss": 5.615242958068848, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.019567854702472687, + "learning_rate": 0.0006, + "loss": 5.5521240234375, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.018347466364502907, + "learning_rate": 0.0006, + "loss": 5.5929107666015625, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.020350053906440735, + "learning_rate": 0.0006, + "loss": 5.6380109786987305, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.016790535300970078, + "learning_rate": 0.0006, + "loss": 5.62994384765625, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.01413649506866932, + "learning_rate": 0.0006, + "loss": 5.6301727294921875, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.013762339949607849, + "learning_rate": 0.0006, + "loss": 5.584138870239258, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.014790690504014492, + "learning_rate": 0.0006, + "loss": 5.666123390197754, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.015471262857317924, + "learning_rate": 0.0006, + "loss": 5.587042808532715, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.015804501250386238, + "learning_rate": 0.0006, + "loss": 5.627497673034668, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.017627960070967674, + "learning_rate": 0.0006, + "loss": 5.556890487670898, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.02074616774916649, + "learning_rate": 0.0006, + "loss": 5.586511135101318, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.026389935985207558, + "learning_rate": 0.0006, + "loss": 5.598520278930664, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.03736288473010063, + "learning_rate": 0.0006, + "loss": 5.575788497924805, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.03523678705096245, + "learning_rate": 0.0006, + "loss": 5.586930751800537, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.024266984313726425, + "learning_rate": 0.0006, + "loss": 5.596694469451904, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.033640433102846146, + "learning_rate": 0.0006, + "loss": 5.536456108093262, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.04689764603972435, + "learning_rate": 0.0006, + "loss": 5.61761999130249, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.04910685867071152, + "learning_rate": 0.0006, + "loss": 5.6298418045043945, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.03152550011873245, + "learning_rate": 0.0006, + "loss": 5.639697074890137, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.037753574550151825, + "learning_rate": 0.0006, + "loss": 5.697381019592285, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.04877670109272003, + "learning_rate": 0.0006, + "loss": 5.605184555053711, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.04228494316339493, + "learning_rate": 0.0006, + "loss": 5.624238967895508, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.03575790300965309, + "learning_rate": 0.0006, + "loss": 5.6037983894348145, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.036425549536943436, + "learning_rate": 0.0006, + "loss": 5.62905216217041, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.04001355543732643, + "learning_rate": 0.0006, + "loss": 5.581656455993652, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.05040939524769783, + "learning_rate": 0.0006, + "loss": 5.621358871459961, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.04140249267220497, + "learning_rate": 0.0006, + "loss": 5.668188095092773, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.033116403967142105, + "learning_rate": 0.0006, + "loss": 5.580477714538574, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.03864264488220215, + "learning_rate": 0.0006, + "loss": 5.539153099060059, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.03026588261127472, + "learning_rate": 0.0006, + "loss": 5.619814872741699, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.02381783165037632, + "learning_rate": 0.0006, + "loss": 5.64422607421875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.02427830547094345, + "learning_rate": 0.0006, + "loss": 5.659873008728027, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.021709900349378586, + "learning_rate": 0.0006, + "loss": 5.590245723724365, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.022812914103269577, + "learning_rate": 0.0006, + "loss": 5.610434532165527, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.019250890240073204, + "learning_rate": 0.0006, + "loss": 5.588078498840332, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.021995501592755318, + "learning_rate": 0.0006, + "loss": 5.541754245758057, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.01477806642651558, + "learning_rate": 0.0006, + "loss": 5.549173355102539, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.016680821776390076, + "learning_rate": 0.0006, + "loss": 5.58221435546875, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.013848318718373775, + "learning_rate": 0.0006, + "loss": 5.5505876541137695, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.016387000679969788, + "learning_rate": 0.0006, + "loss": 5.570460319519043, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.016496233642101288, + "learning_rate": 0.0006, + "loss": 5.513058662414551, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.01675555482506752, + "learning_rate": 0.0006, + "loss": 5.559998989105225, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.016939565539360046, + "learning_rate": 0.0006, + "loss": 5.56727409362793, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.01679506152868271, + "learning_rate": 0.0006, + "loss": 5.603643417358398, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.01656349189579487, + "learning_rate": 0.0006, + "loss": 5.54292631149292, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.01975332945585251, + "learning_rate": 0.0006, + "loss": 5.5225677490234375, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.02329389750957489, + "learning_rate": 0.0006, + "loss": 5.537827491760254, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.026616891846060753, + "learning_rate": 0.0006, + "loss": 5.500011920928955, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.025511575862765312, + "learning_rate": 0.0006, + "loss": 5.5316619873046875, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.025009512901306152, + "learning_rate": 0.0006, + "loss": 5.468493938446045, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.02946239896118641, + "learning_rate": 0.0006, + "loss": 5.5116753578186035, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.036607250571250916, + "learning_rate": 0.0006, + "loss": 5.518636226654053, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.04347814992070198, + "learning_rate": 0.0006, + "loss": 5.572815895080566, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.034463658928871155, + "learning_rate": 0.0006, + "loss": 5.520634651184082, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.03486703336238861, + "learning_rate": 0.0006, + "loss": 5.545441150665283, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.04476739838719368, + "learning_rate": 0.0006, + "loss": 5.554450035095215, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.048745233565568924, + "learning_rate": 0.0006, + "loss": 5.554961204528809, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.04636877775192261, + "learning_rate": 0.0006, + "loss": 5.574655055999756, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.032686877995729446, + "learning_rate": 0.0006, + "loss": 5.5734639167785645, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.03075302764773369, + "learning_rate": 0.0006, + "loss": 5.487662315368652, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.028860175982117653, + "learning_rate": 0.0006, + "loss": 5.519798755645752, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.028599515557289124, + "learning_rate": 0.0006, + "loss": 5.558349609375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.02384260669350624, + "learning_rate": 0.0006, + "loss": 5.520661354064941, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.023180771619081497, + "learning_rate": 0.0006, + "loss": 5.496896743774414, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.021167006343603134, + "learning_rate": 0.0006, + "loss": 5.517192363739014, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02069013938307762, + "learning_rate": 0.0006, + "loss": 5.447656154632568, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 6.161121845245361, + "eval_runtime": 71.5972, + "eval_samples_per_second": 34.107, + "eval_steps_per_second": 1.075, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.022983292117714882, + "learning_rate": 0.0006, + "loss": 5.441346645355225, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.022833511233329773, + "learning_rate": 0.0006, + "loss": 5.45123815536499, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.020015383139252663, + "learning_rate": 0.0006, + "loss": 5.430389404296875, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.016957225278019905, + "learning_rate": 0.0006, + "loss": 5.473883628845215, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.02090892568230629, + "learning_rate": 0.0006, + "loss": 5.467441558837891, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.022806301712989807, + "learning_rate": 0.0006, + "loss": 5.5453081130981445, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.022816946730017662, + "learning_rate": 0.0006, + "loss": 5.426213264465332, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.020761270076036453, + "learning_rate": 0.0006, + "loss": 5.464138507843018, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.01982618309557438, + "learning_rate": 0.0006, + "loss": 5.428033828735352, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.021953724324703217, + "learning_rate": 0.0006, + "loss": 5.456646919250488, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.01871841959655285, + "learning_rate": 0.0006, + "loss": 5.508403778076172, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.01810353808104992, + "learning_rate": 0.0006, + "loss": 5.442957878112793, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.017609186470508575, + "learning_rate": 0.0006, + "loss": 5.4774699211120605, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.01931103691458702, + "learning_rate": 0.0006, + "loss": 5.429748058319092, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.019401391968131065, + "learning_rate": 0.0006, + "loss": 5.466834545135498, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.018862580880522728, + "learning_rate": 0.0006, + "loss": 5.431701183319092, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.01796964928507805, + "learning_rate": 0.0006, + "loss": 5.388613700866699, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.021166296675801277, + "learning_rate": 0.0006, + "loss": 5.454118251800537, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.028568729758262634, + "learning_rate": 0.0006, + "loss": 5.510260581970215, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.03902891278266907, + "learning_rate": 0.0006, + "loss": 5.386396408081055, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.039114370942115784, + "learning_rate": 0.0006, + "loss": 5.457526206970215, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.03560924902558327, + "learning_rate": 0.0006, + "loss": 5.4149980545043945, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.03745810315012932, + "learning_rate": 0.0006, + "loss": 5.464084625244141, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.03383171558380127, + "learning_rate": 0.0006, + "loss": 5.45137882232666, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.027873128652572632, + "learning_rate": 0.0006, + "loss": 5.485784530639648, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.02697703056037426, + "learning_rate": 0.0006, + "loss": 5.459152698516846, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.02646288089454174, + "learning_rate": 0.0006, + "loss": 5.433281421661377, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.02655179798603058, + "learning_rate": 0.0006, + "loss": 5.456627368927002, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.02633797563612461, + "learning_rate": 0.0006, + "loss": 5.484803676605225, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.031187085434794426, + "learning_rate": 0.0006, + "loss": 5.362293243408203, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.02971349097788334, + "learning_rate": 0.0006, + "loss": 5.4369611740112305, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.02996021695435047, + "learning_rate": 0.0006, + "loss": 5.419068813323975, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.032519739121198654, + "learning_rate": 0.0006, + "loss": 5.422189712524414, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.03403490409255028, + "learning_rate": 0.0006, + "loss": 5.470074653625488, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.036989886313676834, + "learning_rate": 0.0006, + "loss": 5.460174560546875, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.03427419438958168, + "learning_rate": 0.0006, + "loss": 5.495920658111572, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.02519620768725872, + "learning_rate": 0.0006, + "loss": 5.474429130554199, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.02473096363246441, + "learning_rate": 0.0006, + "loss": 5.395847320556641, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.02792714536190033, + "learning_rate": 0.0006, + "loss": 5.376262664794922, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.024172596633434296, + "learning_rate": 0.0006, + "loss": 5.459731101989746, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.02304821088910103, + "learning_rate": 0.0006, + "loss": 5.479475498199463, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.02342340536415577, + "learning_rate": 0.0006, + "loss": 5.387955665588379, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.021962853148579597, + "learning_rate": 0.0006, + "loss": 5.385697364807129, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.019193042069673538, + "learning_rate": 0.0006, + "loss": 5.431601047515869, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.021378040313720703, + "learning_rate": 0.0006, + "loss": 5.3070969581604, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.021641355007886887, + "learning_rate": 0.0006, + "loss": 5.360584735870361, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.019899170845746994, + "learning_rate": 0.0006, + "loss": 5.398155212402344, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.02298869751393795, + "learning_rate": 0.0006, + "loss": 5.410561561584473, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.02815551869571209, + "learning_rate": 0.0006, + "loss": 5.298604965209961, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.03041219152510166, + "learning_rate": 0.0006, + "loss": 5.425869941711426, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.025710400193929672, + "learning_rate": 0.0006, + "loss": 5.344316482543945, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.018438996747136116, + "learning_rate": 0.0006, + "loss": 5.379913330078125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.016683703288435936, + "learning_rate": 0.0006, + "loss": 5.338386058807373, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.017252853140234947, + "learning_rate": 0.0006, + "loss": 5.360256195068359, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.019606001675128937, + "learning_rate": 0.0006, + "loss": 5.401415824890137, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.016299517825245857, + "learning_rate": 0.0006, + "loss": 5.39270544052124, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.016657982021570206, + "learning_rate": 0.0006, + "loss": 5.359323978424072, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.017462270334362984, + "learning_rate": 0.0006, + "loss": 5.453655242919922, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.01723778434097767, + "learning_rate": 0.0006, + "loss": 5.454940319061279, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.019071755930781364, + "learning_rate": 0.0006, + "loss": 5.355914115905762, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.02169782482087612, + "learning_rate": 0.0006, + "loss": 5.396806716918945, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.025049131363630295, + "learning_rate": 0.0006, + "loss": 5.347422122955322, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.024977540597319603, + "learning_rate": 0.0006, + "loss": 5.32526969909668, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.024223418906331062, + "learning_rate": 0.0006, + "loss": 5.372445106506348, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.02519913762807846, + "learning_rate": 0.0006, + "loss": 5.280451774597168, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.0255630724132061, + "learning_rate": 0.0006, + "loss": 5.462911605834961, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.022740041837096214, + "learning_rate": 0.0006, + "loss": 5.319089412689209, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.02038649097084999, + "learning_rate": 0.0006, + "loss": 5.37426233291626, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.020509786903858185, + "learning_rate": 0.0006, + "loss": 5.337924957275391, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.021707357838749886, + "learning_rate": 0.0006, + "loss": 5.374853134155273, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.02714136242866516, + "learning_rate": 0.0006, + "loss": 5.2982282638549805, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.027913225814700127, + "learning_rate": 0.0006, + "loss": 5.406044960021973, + "step": 864 + } + ], + "logging_steps": 1, + "max_steps": 28800, + "num_input_tokens_seen": 0, + "num_train_epochs": 400, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.683536033598669e+17, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/training_args.bin b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..c799bcf4051249aabf4c62fff9d38801df7bc2b5 --- /dev/null +++ b/runs/mask15-l2r50-i2-fulle-lm/checkpoint-864/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fb25263250c08c3479a6f0ac9fdf1f196e2a3b08a4eb10384786ff9b1903ef +size 4792